var/home/core/zuul-output/0000755000175000017500000000000015110266527014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110304120015454 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005642105415110304107017674 0ustar rootrootNov 22 08:02:57 crc systemd[1]: Starting Kubernetes Kubelet... Nov 22 08:02:57 crc restorecon[4678]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:57 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 08:02:58 crc restorecon[4678]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 22 08:02:58 crc kubenswrapper[4735]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 08:02:58 crc kubenswrapper[4735]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 22 08:02:58 crc kubenswrapper[4735]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 08:02:58 crc kubenswrapper[4735]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 08:02:58 crc kubenswrapper[4735]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 22 08:02:58 crc kubenswrapper[4735]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.937039 4735 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.964947 4735 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965017 4735 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965029 4735 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965040 4735 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965049 4735 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965058 4735 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965071 4735 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965083 4735 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965092 4735 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965102 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965112 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965121 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965133 4735 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965145 4735 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965156 4735 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965166 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965176 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965186 4735 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965196 4735 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965206 4735 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965229 4735 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965239 4735 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965249 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965259 4735 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965269 4735 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965278 4735 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965287 4735 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965295 4735 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965319 4735 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965329 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965338 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965347 4735 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965356 4735 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965365 4735 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965374 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965385 4735 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965394 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965404 4735 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965413 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965424 4735 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965434 4735 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965443 4735 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965453 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965498 4735 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965510 4735 feature_gate.go:330] unrecognized feature gate: Example Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965520 4735 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965530 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965545 4735 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965555 4735 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965568 4735 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965582 4735 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965595 4735 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965605 4735 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965618 4735 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965629 4735 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965640 4735 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965650 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965661 4735 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965672 4735 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965682 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965693 4735 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965704 4735 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965714 4735 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965724 4735 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965740 4735 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965752 4735 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965766 4735 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965776 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965784 4735 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965795 4735 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.965805 4735 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.966882 4735 flags.go:64] FLAG: --address="0.0.0.0" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.966915 4735 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.966933 4735 flags.go:64] FLAG: --anonymous-auth="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.966945 4735 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.966959 4735 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.966969 4735 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.966981 4735 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.966992 4735 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967003 4735 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967012 4735 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967022 4735 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967031 4735 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967041 4735 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967049 4735 flags.go:64] FLAG: --cgroup-root="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967058 4735 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967067 4735 flags.go:64] FLAG: --client-ca-file="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967076 4735 flags.go:64] FLAG: --cloud-config="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967085 4735 flags.go:64] FLAG: --cloud-provider="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967093 4735 flags.go:64] FLAG: --cluster-dns="[]" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967104 4735 flags.go:64] FLAG: --cluster-domain="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967113 4735 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967122 4735 flags.go:64] FLAG: --config-dir="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967131 4735 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967140 4735 flags.go:64] FLAG: --container-log-max-files="5" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967151 4735 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967160 4735 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967170 4735 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967180 4735 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967189 4735 flags.go:64] FLAG: --contention-profiling="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967199 4735 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967208 4735 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967218 4735 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967226 4735 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967238 4735 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967247 4735 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967255 4735 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967264 4735 flags.go:64] FLAG: --enable-load-reader="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967278 4735 flags.go:64] FLAG: --enable-server="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967287 4735 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967300 4735 flags.go:64] FLAG: --event-burst="100" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967309 4735 flags.go:64] FLAG: --event-qps="50" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967318 4735 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967328 4735 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967338 4735 flags.go:64] FLAG: --eviction-hard="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967349 4735 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967358 4735 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967367 4735 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967376 4735 flags.go:64] FLAG: --eviction-soft="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967386 4735 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967394 4735 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967403 4735 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967412 4735 flags.go:64] FLAG: --experimental-mounter-path="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967421 4735 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967430 4735 flags.go:64] FLAG: --fail-swap-on="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967439 4735 flags.go:64] FLAG: --feature-gates="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967449 4735 flags.go:64] FLAG: --file-check-frequency="20s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967525 4735 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967537 4735 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967546 4735 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967555 4735 flags.go:64] FLAG: --healthz-port="10248" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967565 4735 flags.go:64] FLAG: --help="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967574 4735 flags.go:64] FLAG: --hostname-override="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967583 4735 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967592 4735 flags.go:64] FLAG: --http-check-frequency="20s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967602 4735 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967611 4735 flags.go:64] FLAG: --image-credential-provider-config="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967619 4735 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967628 4735 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967638 4735 flags.go:64] FLAG: --image-service-endpoint="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967646 4735 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967655 4735 flags.go:64] FLAG: --kube-api-burst="100" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967664 4735 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967674 4735 flags.go:64] FLAG: --kube-api-qps="50" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967694 4735 flags.go:64] FLAG: --kube-reserved="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967704 4735 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967713 4735 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967725 4735 flags.go:64] FLAG: --kubelet-cgroups="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967736 4735 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967749 4735 flags.go:64] FLAG: --lock-file="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967760 4735 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967771 4735 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967782 4735 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967806 4735 flags.go:64] FLAG: --log-json-split-stream="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967815 4735 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967824 4735 flags.go:64] FLAG: --log-text-split-stream="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967833 4735 flags.go:64] FLAG: --logging-format="text" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967842 4735 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967851 4735 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967860 4735 flags.go:64] FLAG: --manifest-url="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967870 4735 flags.go:64] FLAG: --manifest-url-header="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967884 4735 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967895 4735 flags.go:64] FLAG: --max-open-files="1000000" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967910 4735 flags.go:64] FLAG: --max-pods="110" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967921 4735 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967934 4735 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967945 4735 flags.go:64] FLAG: --memory-manager-policy="None" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967956 4735 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967968 4735 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967979 4735 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.967990 4735 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968013 4735 flags.go:64] FLAG: --node-status-max-images="50" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968023 4735 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968032 4735 flags.go:64] FLAG: --oom-score-adj="-999" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968041 4735 flags.go:64] FLAG: --pod-cidr="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968050 4735 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968064 4735 flags.go:64] FLAG: --pod-manifest-path="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968073 4735 flags.go:64] FLAG: --pod-max-pids="-1" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968082 4735 flags.go:64] FLAG: --pods-per-core="0" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968091 4735 flags.go:64] FLAG: --port="10250" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968102 4735 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968111 4735 flags.go:64] FLAG: --provider-id="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968119 4735 flags.go:64] FLAG: --qos-reserved="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968128 4735 flags.go:64] FLAG: --read-only-port="10255" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968137 4735 flags.go:64] FLAG: --register-node="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968147 4735 flags.go:64] FLAG: --register-schedulable="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968156 4735 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968173 4735 flags.go:64] FLAG: --registry-burst="10" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968184 4735 flags.go:64] FLAG: --registry-qps="5" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968195 4735 flags.go:64] FLAG: --reserved-cpus="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968206 4735 flags.go:64] FLAG: --reserved-memory="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968220 4735 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968232 4735 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968243 4735 flags.go:64] FLAG: --rotate-certificates="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968255 4735 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968267 4735 flags.go:64] FLAG: --runonce="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968278 4735 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968290 4735 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968300 4735 flags.go:64] FLAG: --seccomp-default="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968309 4735 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968318 4735 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968328 4735 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968337 4735 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968346 4735 flags.go:64] FLAG: --storage-driver-password="root" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968355 4735 flags.go:64] FLAG: --storage-driver-secure="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968364 4735 flags.go:64] FLAG: --storage-driver-table="stats" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968374 4735 flags.go:64] FLAG: --storage-driver-user="root" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968383 4735 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968392 4735 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968402 4735 flags.go:64] FLAG: --system-cgroups="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968410 4735 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968425 4735 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968434 4735 flags.go:64] FLAG: --tls-cert-file="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968443 4735 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968454 4735 flags.go:64] FLAG: --tls-min-version="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968494 4735 flags.go:64] FLAG: --tls-private-key-file="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968505 4735 flags.go:64] FLAG: --topology-manager-policy="none" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968515 4735 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968524 4735 flags.go:64] FLAG: --topology-manager-scope="container" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968533 4735 flags.go:64] FLAG: --v="2" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968544 4735 flags.go:64] FLAG: --version="false" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968555 4735 flags.go:64] FLAG: --vmodule="" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968566 4735 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.968575 4735 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968794 4735 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968807 4735 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968816 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968826 4735 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968835 4735 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968843 4735 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968851 4735 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968858 4735 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968869 4735 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968877 4735 feature_gate.go:330] unrecognized feature gate: Example Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968885 4735 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968893 4735 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968910 4735 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968921 4735 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968931 4735 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968940 4735 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968950 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968958 4735 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968967 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968975 4735 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968984 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.968993 4735 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969002 4735 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969009 4735 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969018 4735 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969026 4735 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969034 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969042 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969051 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969059 4735 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969067 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969075 4735 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969082 4735 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969090 4735 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969098 4735 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969105 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969113 4735 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969121 4735 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969128 4735 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969138 4735 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969148 4735 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969158 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969166 4735 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969175 4735 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969186 4735 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969200 4735 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969210 4735 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969219 4735 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969228 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969236 4735 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969244 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969252 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969260 4735 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969267 4735 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969275 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969283 4735 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969290 4735 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969298 4735 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969306 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969314 4735 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969321 4735 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969329 4735 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969337 4735 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969345 4735 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969354 4735 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969363 4735 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969371 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969378 4735 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969387 4735 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969394 4735 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.969402 4735 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.969427 4735 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.983755 4735 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.983805 4735 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984002 4735 feature_gate.go:330] unrecognized feature gate: Example Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984023 4735 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984037 4735 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984050 4735 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984061 4735 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984074 4735 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984086 4735 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984097 4735 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984108 4735 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984120 4735 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984130 4735 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984141 4735 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984152 4735 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984164 4735 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984175 4735 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984186 4735 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984197 4735 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984208 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984220 4735 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984234 4735 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984246 4735 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984257 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984268 4735 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984284 4735 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984300 4735 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984314 4735 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984329 4735 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984341 4735 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984354 4735 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984365 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984380 4735 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984397 4735 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984410 4735 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984421 4735 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984433 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984445 4735 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984456 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984504 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984515 4735 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984525 4735 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984535 4735 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984546 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984557 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984567 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984578 4735 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984594 4735 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984609 4735 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984626 4735 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984638 4735 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984650 4735 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984661 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984672 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984683 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984694 4735 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984710 4735 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984727 4735 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984739 4735 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984750 4735 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984761 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984771 4735 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984784 4735 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984796 4735 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984806 4735 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984817 4735 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984829 4735 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984840 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984851 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984862 4735 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984873 4735 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984883 4735 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.984895 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.984914 4735 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985230 4735 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985252 4735 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985264 4735 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985276 4735 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985291 4735 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985306 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985318 4735 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985330 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985342 4735 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985354 4735 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985365 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985378 4735 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985389 4735 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985401 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985412 4735 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985422 4735 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985433 4735 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985444 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985455 4735 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985505 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985519 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985530 4735 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985540 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985552 4735 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985570 4735 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985585 4735 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985653 4735 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985670 4735 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985682 4735 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985694 4735 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985706 4735 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985717 4735 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985728 4735 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985739 4735 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985750 4735 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985761 4735 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985772 4735 feature_gate.go:330] unrecognized feature gate: Example Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985783 4735 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985794 4735 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985806 4735 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985817 4735 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985828 4735 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985839 4735 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985851 4735 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985863 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985874 4735 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985885 4735 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985895 4735 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985906 4735 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985918 4735 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985933 4735 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985947 4735 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985958 4735 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985970 4735 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.985981 4735 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986000 4735 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986013 4735 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986027 4735 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986039 4735 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986050 4735 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986061 4735 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986073 4735 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986085 4735 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986098 4735 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986111 4735 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986123 4735 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986136 4735 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986148 4735 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986160 4735 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986172 4735 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 08:02:58 crc kubenswrapper[4735]: W1122 08:02:58.986183 4735 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.986201 4735 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 08:02:58 crc kubenswrapper[4735]: I1122 08:02:58.996435 4735 server.go:940] "Client rotation is on, will bootstrap in background" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.001434 4735 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.001580 4735 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.003423 4735 server.go:997] "Starting client certificate rotation" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.003477 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.003988 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-10 16:06:26.692564776 +0000 UTC Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.004181 4735 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 440h3m27.688390352s for next certificate rotation Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.073357 4735 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.078082 4735 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.110438 4735 log.go:25] "Validated CRI v1 runtime API" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.143406 4735 log.go:25] "Validated CRI v1 image API" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.145806 4735 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.153487 4735 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-22-07-58-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.153747 4735 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.176360 4735 manager.go:217] Machine: {Timestamp:2025-11-22 08:02:59.173874278 +0000 UTC m=+0.778212893 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:461f179d-030f-4c48-b90b-03251dc5d5b9 BootID:2f042384-cd21-4499-89ef-6191a35371a5 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ae:a9:a8 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ae:a9:a8 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:10:5e:73 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f6:7b:82 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:66:b8:44 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:34:be:0a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:f6:26:cc:47:19:98 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fe:c7:a1:a0:97:db Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.176998 4735 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.177223 4735 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.178609 4735 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.178982 4735 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.179105 4735 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.179426 4735 topology_manager.go:138] "Creating topology manager with none policy" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.179717 4735 container_manager_linux.go:303] "Creating device plugin manager" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.180409 4735 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.180615 4735 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.181237 4735 state_mem.go:36] "Initialized new in-memory state store" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.181420 4735 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.184884 4735 kubelet.go:418] "Attempting to sync node with API server" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.185070 4735 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.185184 4735 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.185282 4735 kubelet.go:324] "Adding apiserver pod source" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.185356 4735 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.189969 4735 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.191228 4735 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.193266 4735 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 22 08:02:59 crc kubenswrapper[4735]: W1122 08:02:59.194738 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.194807 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:02:59 crc kubenswrapper[4735]: W1122 08:02:59.194963 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.195056 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.196642 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.196751 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.196815 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.196887 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.196955 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.197021 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.197083 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.197157 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.197221 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.197282 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.197347 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.197408 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.198276 4735 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.200222 4735 server.go:1280] "Started kubelet" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.200296 4735 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.200452 4735 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.200960 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.201184 4735 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 22 08:02:59 crc systemd[1]: Started Kubernetes Kubelet. Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.203924 4735 server.go:460] "Adding debug handlers to kubelet server" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.205110 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.205185 4735 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.205215 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 16:28:51.096465504 +0000 UTC Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.205253 4735 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 200h25m51.891219645s for next certificate rotation Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.205335 4735 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.205380 4735 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.205387 4735 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.205562 4735 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 22 08:02:59 crc kubenswrapper[4735]: W1122 08:02:59.206487 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.206584 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.206703 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="200ms" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.207637 4735 factory.go:55] Registering systemd factory Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.207664 4735 factory.go:221] Registration of the systemd container factory successfully Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.208871 4735 factory.go:153] Registering CRI-O factory Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.208952 4735 factory.go:221] Registration of the crio container factory successfully Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.209097 4735 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.209249 4735 factory.go:103] Registering Raw factory Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.209340 4735 manager.go:1196] Started watching for new ooms in manager Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.210251 4735 manager.go:319] Starting recovery of all containers Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.209092 4735 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a457678907f9b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-22 08:02:59.200163739 +0000 UTC m=+0.804502364,LastTimestamp:2025-11-22 08:02:59.200163739 +0000 UTC m=+0.804502364,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224843 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224893 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224905 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224917 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224927 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224939 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224949 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224957 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224973 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224983 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.224996 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225005 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225017 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225028 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225039 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225047 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225055 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225066 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225074 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225086 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225095 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225104 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225115 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225124 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225136 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225145 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225159 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225170 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225183 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225211 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225223 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225262 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225272 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225284 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225293 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225304 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225314 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225323 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225334 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225344 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225355 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225365 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225374 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225384 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225394 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225406 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225414 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225423 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225434 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225654 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225671 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225686 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225702 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225717 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225729 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225741 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225753 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225761 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225773 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225783 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225796 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225807 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225815 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225828 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225836 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225847 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225855 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225863 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225876 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225884 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225894 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225903 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225912 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225923 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225931 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225942 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225951 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225959 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225970 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225978 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.225991 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226000 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226011 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226021 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226032 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226041 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226052 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226061 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226073 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226081 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226092 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226103 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226111 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226124 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226134 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226143 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226155 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226164 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226175 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226183 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226194 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226204 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226213 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226224 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226245 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226260 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226272 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226283 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226296 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226309 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226320 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226333 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226346 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226356 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226371 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226380 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226401 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226411 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226419 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226430 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.226440 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230753 4735 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230796 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230813 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230828 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230843 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230857 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230870 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230882 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230895 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230915 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230927 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230939 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230951 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230963 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230976 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.230988 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231000 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231012 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231025 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231077 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231126 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231142 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231156 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231170 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231208 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231221 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231234 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231254 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231267 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231279 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231297 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231312 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231327 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231340 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231352 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231380 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231393 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231406 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231421 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231434 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231447 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231477 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231497 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231511 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231525 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231537 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231549 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231562 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231576 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231590 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231602 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231615 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231634 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231649 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231662 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231675 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231691 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231705 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231719 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231732 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231747 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231760 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231774 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231787 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231800 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231814 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231828 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231845 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231858 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231873 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231887 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231901 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231915 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231929 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231943 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231956 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231970 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231983 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.231997 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232010 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232022 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232036 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232049 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232067 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232082 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232094 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232108 4735 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232121 4735 reconstruct.go:97] "Volume reconstruction finished" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.232131 4735 reconciler.go:26] "Reconciler: start to sync state" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.233781 4735 manager.go:324] Recovery completed Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.242989 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.245703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.245813 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.245836 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.248036 4735 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.248073 4735 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.248191 4735 state_mem.go:36] "Initialized new in-memory state store" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.260101 4735 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.262010 4735 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.262051 4735 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.262072 4735 kubelet.go:2335] "Starting kubelet main sync loop" Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.262112 4735 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 22 08:02:59 crc kubenswrapper[4735]: W1122 08:02:59.262575 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.262625 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.263173 4735 policy_none.go:49] "None policy: Start" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.263912 4735 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.263934 4735 state_mem.go:35] "Initializing new in-memory state store" Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.305907 4735 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.319558 4735 manager.go:334] "Starting Device Plugin manager" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.319607 4735 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.319620 4735 server.go:79] "Starting device plugin registration server" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.319962 4735 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.319973 4735 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.320148 4735 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.320242 4735 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.320249 4735 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.326948 4735 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.362870 4735 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.363082 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.365095 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.365160 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.365173 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.365363 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.365509 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.365598 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.366741 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.366776 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.366788 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.366839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.366897 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.366912 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.367282 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.367374 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.367423 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.368611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.368659 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.368673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.368870 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.368950 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.368986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.369000 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.369217 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.369331 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.369583 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.369613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.369625 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.369756 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.369936 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370012 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370582 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370618 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370914 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370947 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370960 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.371004 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.370940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.371068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.371078 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.371621 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.371658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.371667 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.408211 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="400ms" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.420755 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.421974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.422006 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.422033 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.422056 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.422577 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433653 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433706 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433740 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433773 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433803 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433860 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433890 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433940 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433968 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.433995 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.434024 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.434054 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.434134 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.434198 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.434223 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.535415 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.535496 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536039 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536108 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536150 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536268 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536315 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536358 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536183 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536399 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536490 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536399 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536553 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536581 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536598 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536520 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536633 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536649 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536609 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536744 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.536824 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.537011 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.537492 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.537570 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.537631 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.537759 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.537831 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.538387 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.538455 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.538562 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.623647 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.625718 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.625784 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.625798 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.625835 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.626573 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.698919 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.708482 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.729673 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: W1122 08:02:59.750917 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4370e22ef5d9bc27363b132b95fc778e3a3de232c9974ba6d2f40eb3344531e0 WatchSource:0}: Error finding container 4370e22ef5d9bc27363b132b95fc778e3a3de232c9974ba6d2f40eb3344531e0: Status 404 returned error can't find the container with id 4370e22ef5d9bc27363b132b95fc778e3a3de232c9974ba6d2f40eb3344531e0 Nov 22 08:02:59 crc kubenswrapper[4735]: W1122 08:02:59.755700 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-180fcd5f5510311d3abc289a4252ef0727d3a574f43d17c7e058f323095bd02e WatchSource:0}: Error finding container 180fcd5f5510311d3abc289a4252ef0727d3a574f43d17c7e058f323095bd02e: Status 404 returned error can't find the container with id 180fcd5f5510311d3abc289a4252ef0727d3a574f43d17c7e058f323095bd02e Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.756337 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: I1122 08:02:59.765081 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:02:59 crc kubenswrapper[4735]: W1122 08:02:59.775544 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-004b81e671d6a61ca1f7455258350499452ca38fa6cfba672d45e4fe3e982116 WatchSource:0}: Error finding container 004b81e671d6a61ca1f7455258350499452ca38fa6cfba672d45e4fe3e982116: Status 404 returned error can't find the container with id 004b81e671d6a61ca1f7455258350499452ca38fa6cfba672d45e4fe3e982116 Nov 22 08:02:59 crc kubenswrapper[4735]: W1122 08:02:59.793941 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-529c17131bf10ea343adfacc128747def581f0c3db2d4423ceb27f21968c4bf3 WatchSource:0}: Error finding container 529c17131bf10ea343adfacc128747def581f0c3db2d4423ceb27f21968c4bf3: Status 404 returned error can't find the container with id 529c17131bf10ea343adfacc128747def581f0c3db2d4423ceb27f21968c4bf3 Nov 22 08:02:59 crc kubenswrapper[4735]: E1122 08:02:59.809004 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="800ms" Nov 22 08:03:00 crc kubenswrapper[4735]: W1122 08:03:00.013246 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:00 crc kubenswrapper[4735]: E1122 08:03:00.013348 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.027752 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.029408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.029439 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.029450 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.029488 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 08:03:00 crc kubenswrapper[4735]: E1122 08:03:00.029932 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.201880 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.266507 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"529c17131bf10ea343adfacc128747def581f0c3db2d4423ceb27f21968c4bf3"} Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.267571 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"004b81e671d6a61ca1f7455258350499452ca38fa6cfba672d45e4fe3e982116"} Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.268512 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"180fcd5f5510311d3abc289a4252ef0727d3a574f43d17c7e058f323095bd02e"} Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.269545 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c068f0796ba6b56cc244b4f3bd5f653faefcbd567865d8901c8bbda1229439be"} Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.270199 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4370e22ef5d9bc27363b132b95fc778e3a3de232c9974ba6d2f40eb3344531e0"} Nov 22 08:03:00 crc kubenswrapper[4735]: W1122 08:03:00.405209 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:00 crc kubenswrapper[4735]: E1122 08:03:00.405321 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:03:00 crc kubenswrapper[4735]: W1122 08:03:00.575204 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:00 crc kubenswrapper[4735]: E1122 08:03:00.575312 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:03:00 crc kubenswrapper[4735]: E1122 08:03:00.610875 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="1.6s" Nov 22 08:03:00 crc kubenswrapper[4735]: W1122 08:03:00.723897 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:00 crc kubenswrapper[4735]: E1122 08:03:00.724006 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.830732 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.833180 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.833259 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.833274 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:00 crc kubenswrapper[4735]: I1122 08:03:00.833318 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 08:03:00 crc kubenswrapper[4735]: E1122 08:03:00.834045 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.202640 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.275333 4735 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64" exitCode=0 Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.275437 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64"} Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.275586 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.277671 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.277725 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.277742 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.278142 4735 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d6a4d2467c841f0bd41aa23cf9c18b55786412f686d69606c6e947205e3e4d20" exitCode=0 Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.278828 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d6a4d2467c841f0bd41aa23cf9c18b55786412f686d69606c6e947205e3e4d20"} Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.279152 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.281598 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.282564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.282589 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.284511 4735 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088" exitCode=0 Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.284629 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088"} Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.284671 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.285783 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.285822 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.285840 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.287759 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b" exitCode=0 Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.287858 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b"} Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.287871 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.289296 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.289328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.289346 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.290398 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f"} Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.292882 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.293918 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.293963 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:01 crc kubenswrapper[4735]: I1122 08:03:01.293979 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.202548 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:02 crc kubenswrapper[4735]: E1122 08:03:02.212370 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="3.2s" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.299063 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.299151 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.299173 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.299450 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.300967 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.301009 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.301030 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.310597 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.310658 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.310668 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.313794 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.313835 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.313846 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.313960 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.315311 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.315361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.315378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.318658 4735 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92" exitCode=0 Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.318727 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.318857 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.320344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.320373 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.320382 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.321815 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"03bc10ac96df86b5aa93e10f2dfd9153d08eb85a2d8b17efdf729f0360bb5e45"} Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.322075 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.323251 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.323291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.323301 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.434594 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.435934 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.436008 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.436023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:02 crc kubenswrapper[4735]: I1122 08:03:02.436076 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 08:03:02 crc kubenswrapper[4735]: E1122 08:03:02.436713 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Nov 22 08:03:02 crc kubenswrapper[4735]: W1122 08:03:02.711603 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:02 crc kubenswrapper[4735]: E1122 08:03:02.711815 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:03:02 crc kubenswrapper[4735]: W1122 08:03:02.763280 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:02 crc kubenswrapper[4735]: E1122 08:03:02.763395 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:03:03 crc kubenswrapper[4735]: W1122 08:03:03.008877 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:03 crc kubenswrapper[4735]: E1122 08:03:03.008998 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:03:03 crc kubenswrapper[4735]: W1122 08:03:03.072262 4735 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:03 crc kubenswrapper[4735]: E1122 08:03:03.072390 4735 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.202829 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.329208 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963"} Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.329257 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.329278 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6a721229447f0cc68438ac658acec42bf8bb1e743ed5fa48f22f5ee6b9e94d1a"} Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.330565 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.330617 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.330630 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.333406 4735 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721" exitCode=0 Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.333452 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721"} Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.333518 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.333662 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.333676 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.333804 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.334093 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.334422 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.334445 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.334524 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.334924 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.334947 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.334960 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.335136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.335205 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.335233 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.336861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.336934 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.336953 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:03 crc kubenswrapper[4735]: I1122 08:03:03.726516 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.343269 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6"} Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.343333 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.343351 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720"} Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.343372 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5"} Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.343385 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81"} Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.343391 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.344336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.344401 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:04 crc kubenswrapper[4735]: I1122 08:03:04.344420 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.350158 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda"} Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.350202 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.350250 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.350323 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.351261 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.351294 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.351306 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.351986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.352027 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.352040 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.637409 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.639056 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.639100 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.639112 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:05 crc kubenswrapper[4735]: I1122 08:03:05.639149 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.353055 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.354263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.354312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.354326 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.436615 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.436890 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.438631 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.438701 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.438713 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.852044 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.852296 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.853435 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.853541 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:06 crc kubenswrapper[4735]: I1122 08:03:06.853568 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.155572 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.155833 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.157157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.157213 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.157226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.218432 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.218700 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.219929 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.219976 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:08 crc kubenswrapper[4735]: I1122 08:03:08.219985 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.026496 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.026635 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.031088 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.031123 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.031134 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.037049 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:03:09 crc kubenswrapper[4735]: E1122 08:03:09.327142 4735 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.360739 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.360804 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.361750 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.361798 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.361811 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:09 crc kubenswrapper[4735]: I1122 08:03:09.737344 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.114591 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.114842 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.116357 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.116531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.116567 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.362691 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.364540 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.364684 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.364702 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.367255 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.623919 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.624163 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.625607 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.625648 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:10 crc kubenswrapper[4735]: I1122 08:03:10.625657 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:11 crc kubenswrapper[4735]: I1122 08:03:11.365671 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:11 crc kubenswrapper[4735]: I1122 08:03:11.366869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:11 crc kubenswrapper[4735]: I1122 08:03:11.366959 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:11 crc kubenswrapper[4735]: I1122 08:03:11.366986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:12 crc kubenswrapper[4735]: I1122 08:03:12.368158 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:12 crc kubenswrapper[4735]: I1122 08:03:12.369271 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:12 crc kubenswrapper[4735]: I1122 08:03:12.369324 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:12 crc kubenswrapper[4735]: I1122 08:03:12.369344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:12 crc kubenswrapper[4735]: I1122 08:03:12.737940 4735 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 08:03:12 crc kubenswrapper[4735]: I1122 08:03:12.738082 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.203546 4735 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.377653 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.380169 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6a721229447f0cc68438ac658acec42bf8bb1e743ed5fa48f22f5ee6b9e94d1a" exitCode=255 Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.380223 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6a721229447f0cc68438ac658acec42bf8bb1e743ed5fa48f22f5ee6b9e94d1a"} Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.380416 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.381446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.381534 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.381553 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.382375 4735 scope.go:117] "RemoveContainer" containerID="6a721229447f0cc68438ac658acec42bf8bb1e743ed5fa48f22f5ee6b9e94d1a" Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.402624 4735 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.402777 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.408552 4735 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 22 08:03:14 crc kubenswrapper[4735]: I1122 08:03:14.408631 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 22 08:03:15 crc kubenswrapper[4735]: I1122 08:03:15.385741 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 08:03:15 crc kubenswrapper[4735]: I1122 08:03:15.387776 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37"} Nov 22 08:03:15 crc kubenswrapper[4735]: I1122 08:03:15.387981 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:15 crc kubenswrapper[4735]: I1122 08:03:15.388856 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:15 crc kubenswrapper[4735]: I1122 08:03:15.388907 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:15 crc kubenswrapper[4735]: I1122 08:03:15.388919 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:16 crc kubenswrapper[4735]: I1122 08:03:16.853078 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:16 crc kubenswrapper[4735]: I1122 08:03:16.853402 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:16 crc kubenswrapper[4735]: I1122 08:03:16.855392 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:16 crc kubenswrapper[4735]: I1122 08:03:16.855448 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:16 crc kubenswrapper[4735]: I1122 08:03:16.855505 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.231532 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.231733 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.233191 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.233237 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.233259 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.236829 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.397975 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.399351 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.399424 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:18 crc kubenswrapper[4735]: I1122 08:03:18.399442 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:19 crc kubenswrapper[4735]: E1122 08:03:19.327260 4735 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 08:03:19 crc kubenswrapper[4735]: E1122 08:03:19.401927 4735 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.404342 4735 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.405443 4735 trace.go:236] Trace[476671066]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 08:03:07.370) (total time: 12034ms): Nov 22 08:03:19 crc kubenswrapper[4735]: Trace[476671066]: ---"Objects listed" error: 12034ms (08:03:19.405) Nov 22 08:03:19 crc kubenswrapper[4735]: Trace[476671066]: [12.034382381s] [12.034382381s] END Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.405500 4735 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.406719 4735 trace.go:236] Trace[533682737]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 08:03:06.283) (total time: 13122ms): Nov 22 08:03:19 crc kubenswrapper[4735]: Trace[533682737]: ---"Objects listed" error: 13122ms (08:03:19.406) Nov 22 08:03:19 crc kubenswrapper[4735]: Trace[533682737]: [13.122931871s] [13.122931871s] END Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.406845 4735 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.411676 4735 trace.go:236] Trace[1746439891]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 08:03:07.829) (total time: 11582ms): Nov 22 08:03:19 crc kubenswrapper[4735]: Trace[1746439891]: ---"Objects listed" error: 11582ms (08:03:19.411) Nov 22 08:03:19 crc kubenswrapper[4735]: Trace[1746439891]: [11.582121764s] [11.582121764s] END Nov 22 08:03:19 crc kubenswrapper[4735]: E1122 08:03:19.411718 4735 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.411734 4735 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.449216 4735 trace.go:236] Trace[1119548771]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 08:03:07.205) (total time: 12243ms): Nov 22 08:03:19 crc kubenswrapper[4735]: Trace[1119548771]: ---"Objects listed" error: 12243ms (08:03:19.449) Nov 22 08:03:19 crc kubenswrapper[4735]: Trace[1119548771]: [12.243362758s] [12.243362758s] END Nov 22 08:03:19 crc kubenswrapper[4735]: I1122 08:03:19.449277 4735 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.198188 4735 apiserver.go:52] "Watching apiserver" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.201232 4735 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.201513 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.201987 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.202173 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.202207 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.202294 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.202361 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.202420 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.202482 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.202496 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.202423 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.205239 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.206450 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.206675 4735 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.209353 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.209667 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.209834 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210023 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210193 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210324 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210579 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210620 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210644 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210668 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210693 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210714 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210736 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210759 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210780 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210805 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210827 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210848 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210870 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210894 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210919 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210942 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210963 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.210987 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211010 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211057 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211082 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211123 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211161 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211187 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211210 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211246 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211267 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211289 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211312 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211336 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211360 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211379 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211400 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211423 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211450 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211490 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211515 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211538 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211610 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211634 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211659 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211684 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211709 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211736 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211760 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211784 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211806 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211833 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211856 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211880 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211902 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211923 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211944 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211966 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.211988 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212010 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212034 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212075 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212099 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212123 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212146 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212169 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212191 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212213 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212235 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212257 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212277 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212297 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212322 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212347 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212370 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212394 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212417 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212439 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212473 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212496 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212521 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212546 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212569 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212591 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212611 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212633 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212654 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212676 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212697 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212718 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212745 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212766 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212787 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212811 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212833 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212887 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212912 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212936 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212959 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.212980 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213001 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213023 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213046 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213068 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213094 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213116 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213137 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213159 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213180 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213204 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213227 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213251 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213274 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213298 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213362 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213385 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213410 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213433 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213487 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213510 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213534 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213557 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213590 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213615 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213639 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213664 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213687 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213710 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213743 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213768 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213793 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213820 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213844 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213867 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213893 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213917 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213943 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213968 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.213993 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214019 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214050 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214074 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214099 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214123 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214148 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214171 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214195 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214219 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214243 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214267 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214292 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214317 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214345 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214370 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214396 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214420 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214449 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214496 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214522 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214547 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214572 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214595 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214619 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214644 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214670 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214696 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214720 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214744 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214770 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214889 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214918 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214942 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214966 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.214989 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215014 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215041 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215069 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215093 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215118 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215142 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215167 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215193 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215235 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215260 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215284 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215313 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215339 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215367 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215394 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215419 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215444 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215487 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215513 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215538 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215562 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215589 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215617 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215644 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215669 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215730 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215760 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215786 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215814 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215840 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215867 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215900 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215931 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215959 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.215985 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.216009 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.216038 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.216067 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.216093 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.216180 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.216242 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:20.716222188 +0000 UTC m=+22.320560803 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.216751 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.216905 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.216951 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.217192 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.217478 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.217703 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.217790 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.217790 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.217975 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218115 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218133 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218399 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218409 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218632 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218710 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218732 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218779 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.218977 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.219057 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.219281 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.219423 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.219656 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.219798 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.219815 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.219821 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.220201 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.220512 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.220543 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.220670 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.220898 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221108 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221362 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221583 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221659 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221713 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221848 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221892 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221911 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.221939 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222025 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222057 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222136 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222325 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222348 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222372 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222377 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222542 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222613 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.222870 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223013 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223148 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223253 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223342 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223215 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223547 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223553 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223635 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223642 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.223838 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.225624 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.225665 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.225886 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.226073 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.226085 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.224026 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.224071 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.225058 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.225356 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.226630 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.227678 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.227758 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.227805 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.228031 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.228351 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.228393 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.228522 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.228814 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.228865 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.228940 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.229259 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.229382 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.229762 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.230164 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.230712 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.231223 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.231921 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.232238 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.233087 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.233571 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.233876 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.224030 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.236502 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.237705 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.237964 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.238350 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.238709 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.238807 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.239191 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.239567 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.239615 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.239824 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.239880 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.239203 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.240828 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.241884 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.242351 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.242362 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.242588 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.242906 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.242904 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.243017 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.243215 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.243742 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.243940 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.243762 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.244596 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.244880 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.244880 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.245284 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.246123 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.245876 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.246303 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.246377 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.246426 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.246427 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.249673 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.249689 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.250166 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.250440 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.250565 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.250581 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.250957 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.250773 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.251707 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.251774 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.251821 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.252028 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.252025 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:03:20.751914398 +0000 UTC m=+22.356253013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.252374 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.253402 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.253887 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.255921 4735 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.255614 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.254637 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.252703 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.254926 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.254996 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.255099 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.255094 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.262953 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263022 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263223 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263298 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263339 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263410 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263476 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263583 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263789 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263813 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.263841 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:20.755723739 +0000 UTC m=+22.360062364 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263843 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263872 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.263901 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.264016 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.264015 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.255108 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.255440 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.256746 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.256779 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.256889 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.257123 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.257117 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.257155 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.260081 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.260774 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.260967 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.261080 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.261303 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.261411 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.261519 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.264612 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.261554 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.262670 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.262736 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.265187 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.265285 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.265822 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.266056 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.272006 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.274189 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.279039 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.279615 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.279590 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.280625 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.282654 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.284992 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.285533 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.285919 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.286023 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.286103 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.286245 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:20.78621636 +0000 UTC m=+22.390554985 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.289135 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.289119 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.290821 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.291392 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.290137 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.290025 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.291445 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.291649 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.298529 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.298776 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.298840 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.298965 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:20.798941359 +0000 UTC m=+22.403279964 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.298316 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.299033 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.302179 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.303209 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.317851 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.317926 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.317988 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318077 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318091 4735 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318102 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318120 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318133 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318145 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318157 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318169 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318181 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318195 4735 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318208 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318220 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318231 4735 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318259 4735 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318309 4735 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318327 4735 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318339 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318352 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318364 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318375 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318387 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318398 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318410 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318423 4735 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318435 4735 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318451 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318481 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318493 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318507 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318518 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318530 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318541 4735 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318554 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318566 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318859 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.318101 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.319299 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.319977 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.319997 4735 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320009 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320021 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320031 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.319852 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320041 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320093 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320107 4735 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320121 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320135 4735 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320147 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320159 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320172 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320183 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320195 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320207 4735 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320218 4735 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320230 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320242 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320253 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320264 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320275 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320285 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320296 4735 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320307 4735 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320319 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320329 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320339 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320350 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320361 4735 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320372 4735 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320384 4735 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320396 4735 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320409 4735 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320421 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320433 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320445 4735 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320502 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320516 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320528 4735 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320539 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320550 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320562 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320573 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320585 4735 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320596 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320607 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320618 4735 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320629 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320916 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320937 4735 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320950 4735 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320962 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320973 4735 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320984 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.320995 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321006 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321017 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321029 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321040 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321051 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321063 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321074 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321086 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321546 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321565 4735 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321589 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321603 4735 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321615 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321626 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321637 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321673 4735 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321688 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321701 4735 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321711 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321723 4735 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321735 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321747 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321759 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321771 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321784 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321797 4735 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321812 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: W1122 08:03:20.321803 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-0dafe380a5d992c3204c43153ceaeacec1e070112b9b00163b592c3b44e8bda0 WatchSource:0}: Error finding container 0dafe380a5d992c3204c43153ceaeacec1e070112b9b00163b592c3b44e8bda0: Status 404 returned error can't find the container with id 0dafe380a5d992c3204c43153ceaeacec1e070112b9b00163b592c3b44e8bda0 Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321826 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321888 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321901 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321914 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321924 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321932 4735 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321947 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321958 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321968 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.321977 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322001 4735 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322013 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322022 4735 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322033 4735 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322044 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322055 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322123 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322137 4735 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322153 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322169 4735 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322182 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322196 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322208 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322225 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322236 4735 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322251 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322265 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322279 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322288 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322299 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322309 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322318 4735 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322328 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322337 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322193 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322347 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322575 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322640 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322720 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322742 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322755 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322768 4735 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322780 4735 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322792 4735 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322805 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322818 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322829 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322843 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322855 4735 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322866 4735 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322892 4735 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322906 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322919 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322931 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322944 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322956 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322968 4735 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322979 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.322992 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323015 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323026 4735 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323039 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323051 4735 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323062 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323074 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323085 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323097 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323222 4735 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323234 4735 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323247 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323284 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323296 4735 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323309 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323641 4735 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323658 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.323894 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.335332 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.336089 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.351320 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.369904 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.381859 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.420243 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.427711 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d0e5addd526048a8aa0a1161f45027d223907d44d393f9bb32898caa2fe58573"} Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.428096 4735 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.428118 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.428131 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.433477 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0dafe380a5d992c3204c43153ceaeacec1e070112b9b00163b592c3b44e8bda0"} Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.563975 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 08:03:20 crc kubenswrapper[4735]: W1122 08:03:20.576240 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-06bf5e8d03e1727647671ca0898049ce30dd367209f2cc3033edfab366b5be18 WatchSource:0}: Error finding container 06bf5e8d03e1727647671ca0898049ce30dd367209f2cc3033edfab366b5be18: Status 404 returned error can't find the container with id 06bf5e8d03e1727647671ca0898049ce30dd367209f2cc3033edfab366b5be18 Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.657746 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.670752 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.678634 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.685881 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.720695 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.729945 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.730081 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.730180 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:21.730159867 +0000 UTC m=+23.334498532 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.748036 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.777715 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.779691 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.805250 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.823047 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.830842 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.830933 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.830961 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.830991 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831076 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:03:21.83104903 +0000 UTC m=+23.435387625 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831153 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831184 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831196 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831208 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831224 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831238 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831256 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:21.831237186 +0000 UTC m=+23.435575791 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831277 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:21.831269427 +0000 UTC m=+23.435608032 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831330 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: E1122 08:03:20.831353 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:21.831347629 +0000 UTC m=+23.435686234 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.834937 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.845981 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.862642 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.875614 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.886742 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.899262 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.915579 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:20 crc kubenswrapper[4735]: I1122 08:03:20.925946 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.266841 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.267324 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.268051 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.268624 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.269142 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.269610 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.270161 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.270715 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.271285 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.271848 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.272417 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.273189 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.273764 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.274421 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.275101 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.278740 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.279409 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.279883 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.281198 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.281903 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.282446 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.283642 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.284165 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.285382 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.285913 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.287155 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.287953 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.289017 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.289751 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.290584 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.291033 4735 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.291132 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.292791 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.293614 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.294006 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.295453 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.296563 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.297036 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.298304 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.299121 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.299587 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.300154 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.300748 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.301309 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.301799 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.302299 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.302858 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.303558 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.304008 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.304448 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.304911 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.305407 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.306009 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.306450 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.309771 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-x2srr"] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.310099 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-x2srr" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.312336 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.312729 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.312929 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.323218 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.334778 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64hff\" (UniqueName: \"kubernetes.io/projected/78c491a2-c7e5-4084-a48b-c6dd8163b467-kube-api-access-64hff\") pod \"node-resolver-x2srr\" (UID: \"78c491a2-c7e5-4084-a48b-c6dd8163b467\") " pod="openshift-dns/node-resolver-x2srr" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.334850 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/78c491a2-c7e5-4084-a48b-c6dd8163b467-hosts-file\") pod \"node-resolver-x2srr\" (UID: \"78c491a2-c7e5-4084-a48b-c6dd8163b467\") " pod="openshift-dns/node-resolver-x2srr" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.335663 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.348703 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.364191 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.378000 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.398836 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.414494 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.435219 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/78c491a2-c7e5-4084-a48b-c6dd8163b467-hosts-file\") pod \"node-resolver-x2srr\" (UID: \"78c491a2-c7e5-4084-a48b-c6dd8163b467\") " pod="openshift-dns/node-resolver-x2srr" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.435287 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64hff\" (UniqueName: \"kubernetes.io/projected/78c491a2-c7e5-4084-a48b-c6dd8163b467-kube-api-access-64hff\") pod \"node-resolver-x2srr\" (UID: \"78c491a2-c7e5-4084-a48b-c6dd8163b467\") " pod="openshift-dns/node-resolver-x2srr" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.435352 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/78c491a2-c7e5-4084-a48b-c6dd8163b467-hosts-file\") pod \"node-resolver-x2srr\" (UID: \"78c491a2-c7e5-4084-a48b-c6dd8163b467\") " pod="openshift-dns/node-resolver-x2srr" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.437941 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d"} Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.438039 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"06bf5e8d03e1727647671ca0898049ce30dd367209f2cc3033edfab366b5be18"} Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.440965 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.441706 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.444204 4735 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37" exitCode=255 Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.444312 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37"} Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.444433 4735 scope.go:117] "RemoveContainer" containerID="6a721229447f0cc68438ac658acec42bf8bb1e743ed5fa48f22f5ee6b9e94d1a" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.446938 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37"} Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.447005 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e"} Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.450806 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.463704 4735 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.466385 4735 scope.go:117] "RemoveContainer" containerID="33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.466590 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.466696 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.470018 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64hff\" (UniqueName: \"kubernetes.io/projected/78c491a2-c7e5-4084-a48b-c6dd8163b467-kube-api-access-64hff\") pod \"node-resolver-x2srr\" (UID: \"78c491a2-c7e5-4084-a48b-c6dd8163b467\") " pod="openshift-dns/node-resolver-x2srr" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.478519 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.508844 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.533753 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.564165 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.622664 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.627167 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-x2srr" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.638447 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: W1122 08:03:21.650731 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78c491a2_c7e5_4084_a48b_c6dd8163b467.slice/crio-9e96bd32e7f659cd0f99694ed676931fd1677b63f0fce1782c9306adcb677c51 WatchSource:0}: Error finding container 9e96bd32e7f659cd0f99694ed676931fd1677b63f0fce1782c9306adcb677c51: Status 404 returned error can't find the container with id 9e96bd32e7f659cd0f99694ed676931fd1677b63f0fce1782c9306adcb677c51 Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.656174 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.680292 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.684120 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-dt2lg"] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.684709 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.686333 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.686531 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.686724 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-85z2h"] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.687005 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.687199 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.687256 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.688070 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.689771 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.689984 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.690710 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pl875"] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.691631 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.692865 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-25pl8"] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.693544 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.693882 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.694178 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.694333 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.695048 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.695182 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.698357 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.698558 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.699716 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.699823 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.700038 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.700038 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.700343 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.700837 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.714774 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.732528 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a721229447f0cc68438ac658acec42bf8bb1e743ed5fa48f22f5ee6b9e94d1a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:13Z\\\",\\\"message\\\":\\\"W1122 08:03:02.968281 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1122 08:03:02.969083 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763798582 cert, and key in /tmp/serving-cert-3145015104/serving-signer.crt, /tmp/serving-cert-3145015104/serving-signer.key\\\\nI1122 08:03:03.377353 1 observer_polling.go:159] Starting file observer\\\\nW1122 08:03:03.380108 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 08:03:03.380357 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:03.381355 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3145015104/tls.crt::/tmp/serving-cert-3145015104/tls.key\\\\\\\"\\\\nF1122 08:03:13.677856 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.738729 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5003f416-41bd-48b2-8965-292d0f933500-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.738777 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-env-overrides\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.738836 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-netns\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.738860 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-cni-multus\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.738990 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-hostroot\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739064 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-socket-dir-parent\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739101 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-netns\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739129 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739212 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2719040d-3088-41c7-8f16-5508d78669e2-ovn-node-metrics-cert\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739249 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739279 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgldm\" (UniqueName: \"kubernetes.io/projected/5003f416-41bd-48b2-8965-292d0f933500-kube-api-access-lgldm\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739307 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-daemon-config\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739331 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-tuning-conf-dir\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739358 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-systemd-units\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.739359 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739381 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-bin\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739452 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7945a41-afa2-43b7-95a2-085e3b62e64b-mcd-auth-proxy-config\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.739486 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:23.739443991 +0000 UTC m=+25.343782776 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739521 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvx5h\" (UniqueName: \"kubernetes.io/projected/2719040d-3088-41c7-8f16-5508d78669e2-kube-api-access-dvx5h\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739595 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-conf-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739663 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-var-lib-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739711 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-cnibin\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739741 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-k8s-cni-cncf-io\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739767 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-cni-bin\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739789 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-ovn-kubernetes\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739840 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-systemd\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739864 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-netd\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739886 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-os-release\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739910 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-cnibin\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739935 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-etc-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739963 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-node-log\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.739990 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-kubelet\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740011 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-config\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740052 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7945a41-afa2-43b7-95a2-085e3b62e64b-proxy-tls\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740085 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t88z4\" (UniqueName: \"kubernetes.io/projected/b7945a41-afa2-43b7-95a2-085e3b62e64b-kube-api-access-t88z4\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740110 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-system-cni-dir\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740140 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-system-cni-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740165 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-etc-kubernetes\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740253 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740324 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-slash\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740354 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-log-socket\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740380 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-script-lib\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740402 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-kubelet\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740493 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz5mx\" (UniqueName: \"kubernetes.io/projected/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-kube-api-access-sz5mx\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740555 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-os-release\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740583 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5003f416-41bd-48b2-8965-292d0f933500-cni-binary-copy\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740615 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7945a41-afa2-43b7-95a2-085e3b62e64b-rootfs\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-cni-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740667 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-cni-binary-copy\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740687 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-multus-certs\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.740710 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-ovn\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.751126 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a721229447f0cc68438ac658acec42bf8bb1e743ed5fa48f22f5ee6b9e94d1a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:13Z\\\",\\\"message\\\":\\\"W1122 08:03:02.968281 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1122 08:03:02.969083 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763798582 cert, and key in /tmp/serving-cert-3145015104/serving-signer.crt, /tmp/serving-cert-3145015104/serving-signer.key\\\\nI1122 08:03:03.377353 1 observer_polling.go:159] Starting file observer\\\\nW1122 08:03:03.380108 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 08:03:03.380357 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:03.381355 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3145015104/tls.crt::/tmp/serving-cert-3145015104/tls.key\\\\\\\"\\\\nF1122 08:03:13.677856 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.773657 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.788502 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.810034 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.823173 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.841530 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.841654 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.841687 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.841719 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.841772 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.841785 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:03:23.84173164 +0000 UTC m=+25.446070265 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.841864 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-slash\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.841874 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.841919 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.841915 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.841959 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.841929 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-kubelet\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842005 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz5mx\" (UniqueName: \"kubernetes.io/projected/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-kube-api-access-sz5mx\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.841975 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842020 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-kubelet\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842032 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-os-release\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.841938 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842112 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-slash\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842217 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-os-release\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.842103 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:23.842078 +0000 UTC m=+25.446416805 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842312 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5003f416-41bd-48b2-8965-292d0f933500-cni-binary-copy\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842373 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-log-socket\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842418 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-script-lib\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.842441 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:23.842412739 +0000 UTC m=+25.446751534 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842504 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7945a41-afa2-43b7-95a2-085e3b62e64b-rootfs\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842522 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-log-socket\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842548 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-cni-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842586 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7945a41-afa2-43b7-95a2-085e3b62e64b-rootfs\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842591 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-cni-binary-copy\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842630 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-multus-certs\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842667 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-ovn\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842701 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-netns\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842741 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-cni-multus\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842776 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-hostroot\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842810 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5003f416-41bd-48b2-8965-292d0f933500-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842847 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-cni-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842860 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-env-overrides\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842894 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-multus-certs\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842925 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842936 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-ovn\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.842968 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-socket-dir-parent\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843006 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-netns\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843046 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843036 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-netns\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843103 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgldm\" (UniqueName: \"kubernetes.io/projected/5003f416-41bd-48b2-8965-292d0f933500-kube-api-access-lgldm\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843193 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2719040d-3088-41c7-8f16-5508d78669e2-ovn-node-metrics-cert\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843247 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7945a41-afa2-43b7-95a2-085e3b62e64b-mcd-auth-proxy-config\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843274 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-daemon-config\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843268 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-cni-multus\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843330 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-script-lib\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843299 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-tuning-conf-dir\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843409 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-systemd-units\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843431 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-hostroot\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843488 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-bin\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843440 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-bin\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843532 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-systemd-units\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843575 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvx5h\" (UniqueName: \"kubernetes.io/projected/2719040d-3088-41c7-8f16-5508d78669e2-kube-api-access-dvx5h\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843596 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-socket-dir-parent\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843613 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-cnibin\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843674 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-conf-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843702 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-var-lib-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843711 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5003f416-41bd-48b2-8965-292d0f933500-cni-binary-copy\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843736 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-cni-binary-copy\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843728 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-k8s-cni-cncf-io\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843787 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-run-k8s-cni-cncf-io\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843788 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-cni-bin\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843852 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-netns\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.843880 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843920 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843880 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-ovn-kubernetes\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: E1122 08:03:21.843932 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:23.843920859 +0000 UTC m=+25.448259664 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843856 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-ovn-kubernetes\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843982 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-conf-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844003 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-systemd\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844032 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-os-release\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844055 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-cnibin\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844076 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-etc-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844086 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-cnibin\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844099 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-node-log\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844123 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-netd\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844138 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-var-lib-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844146 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-kubelet\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844174 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-kubelet\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844186 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-env-overrides\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844204 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7945a41-afa2-43b7-95a2-085e3b62e64b-proxy-tls\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844217 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-etc-openvswitch\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844211 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-cnibin\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844236 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-node-log\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844229 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5003f416-41bd-48b2-8965-292d0f933500-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844247 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t88z4\" (UniqueName: \"kubernetes.io/projected/b7945a41-afa2-43b7-95a2-085e3b62e64b-kube-api-access-t88z4\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844261 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-systemd\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.843813 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-host-var-lib-cni-bin\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844268 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-netd\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844292 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-config\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844306 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-os-release\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844335 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-system-cni-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844370 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-etc-kubernetes\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844389 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-multus-daemon-config\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844402 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-system-cni-dir\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844449 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-etc-kubernetes\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844485 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-system-cni-dir\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844529 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7945a41-afa2-43b7-95a2-085e3b62e64b-mcd-auth-proxy-config\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844571 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-system-cni-dir\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844573 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5003f416-41bd-48b2-8965-292d0f933500-tuning-conf-dir\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.844843 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-config\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.852203 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.853770 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7945a41-afa2-43b7-95a2-085e3b62e64b-proxy-tls\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.854359 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2719040d-3088-41c7-8f16-5508d78669e2-ovn-node-metrics-cert\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.869529 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.870955 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvx5h\" (UniqueName: \"kubernetes.io/projected/2719040d-3088-41c7-8f16-5508d78669e2-kube-api-access-dvx5h\") pod \"ovnkube-node-pl875\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.871033 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz5mx\" (UniqueName: \"kubernetes.io/projected/61dbab7f-c903-4b55-9d90-aacfd12f7f0f-kube-api-access-sz5mx\") pod \"multus-85z2h\" (UID: \"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\") " pod="openshift-multus/multus-85z2h" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.871407 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgldm\" (UniqueName: \"kubernetes.io/projected/5003f416-41bd-48b2-8965-292d0f933500-kube-api-access-lgldm\") pod \"multus-additional-cni-plugins-25pl8\" (UID: \"5003f416-41bd-48b2-8965-292d0f933500\") " pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.872371 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t88z4\" (UniqueName: \"kubernetes.io/projected/b7945a41-afa2-43b7-95a2-085e3b62e64b-kube-api-access-t88z4\") pod \"machine-config-daemon-dt2lg\" (UID: \"b7945a41-afa2-43b7-95a2-085e3b62e64b\") " pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.881867 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.903327 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.915391 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.931209 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.947442 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.960351 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:21 crc kubenswrapper[4735]: I1122 08:03:21.982505 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:21Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.002070 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.013591 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-85z2h" Nov 22 08:03:22 crc kubenswrapper[4735]: W1122 08:03:22.017109 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7945a41_afa2_43b7_95a2_085e3b62e64b.slice/crio-5256e5f72ddb41af07d93906b65b9524817842595d50091e8f64d9fbec09c7a1 WatchSource:0}: Error finding container 5256e5f72ddb41af07d93906b65b9524817842595d50091e8f64d9fbec09c7a1: Status 404 returned error can't find the container with id 5256e5f72ddb41af07d93906b65b9524817842595d50091e8f64d9fbec09c7a1 Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.020630 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.029385 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-25pl8" Nov 22 08:03:22 crc kubenswrapper[4735]: W1122 08:03:22.070651 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5003f416_41bd_48b2_8965_292d0f933500.slice/crio-679a501b490dab1625a16f621c35ce37b776d48b226777b596c19a75a5206215 WatchSource:0}: Error finding container 679a501b490dab1625a16f621c35ce37b776d48b226777b596c19a75a5206215: Status 404 returned error can't find the container with id 679a501b490dab1625a16f621c35ce37b776d48b226777b596c19a75a5206215 Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.263162 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:22 crc kubenswrapper[4735]: E1122 08:03:22.263317 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.263398 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:22 crc kubenswrapper[4735]: E1122 08:03:22.263494 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.263555 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:22 crc kubenswrapper[4735]: E1122 08:03:22.263667 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.450550 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-85z2h" event={"ID":"61dbab7f-c903-4b55-9d90-aacfd12f7f0f","Type":"ContainerStarted","Data":"04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.450609 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-85z2h" event={"ID":"61dbab7f-c903-4b55-9d90-aacfd12f7f0f","Type":"ContainerStarted","Data":"95badbba38378f6114c07e0ee15fd01c6e9b0b6a9cd0515bbfd7b7ddc052d745"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.451952 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.455889 4735 scope.go:117] "RemoveContainer" containerID="33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37" Nov 22 08:03:22 crc kubenswrapper[4735]: E1122 08:03:22.456082 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.456444 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerStarted","Data":"679a501b490dab1625a16f621c35ce37b776d48b226777b596c19a75a5206215"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.457991 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.458019 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"1f0b9a794a58e1b354d17e739c2c69508e496ff0bab41382b9a2c8a4ac474046"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.461604 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.461662 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"5256e5f72ddb41af07d93906b65b9524817842595d50091e8f64d9fbec09c7a1"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.464314 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-x2srr" event={"ID":"78c491a2-c7e5-4084-a48b-c6dd8163b467","Type":"ContainerStarted","Data":"2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.464349 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-x2srr" event={"ID":"78c491a2-c7e5-4084-a48b-c6dd8163b467","Type":"ContainerStarted","Data":"9e96bd32e7f659cd0f99694ed676931fd1677b63f0fce1782c9306adcb677c51"} Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.478049 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.491742 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.505326 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.526632 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.538794 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.556760 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.569691 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.583996 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.603253 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.629012 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.653365 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.674419 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.692503 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.717602 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.747103 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.758150 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.771292 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.792029 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.805050 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.827129 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.842507 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.856078 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.869283 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.880178 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.894094 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.905736 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.917749 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:22 crc kubenswrapper[4735]: I1122 08:03:22.929715 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:22Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.467858 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f"} Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.470840 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerStarted","Data":"de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887"} Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.473597 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156" exitCode=0 Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.473688 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156"} Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.482383 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1"} Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.490547 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.504231 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.526878 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.546447 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.562303 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.584536 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.601941 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.616501 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.631722 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.648011 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.667432 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.691179 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.702997 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.734859 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.748179 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.763916 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.766778 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.766999 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.767124 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:27.767097389 +0000 UTC m=+29.371436034 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.778258 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.791404 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.810936 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.831063 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.846342 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.878311 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.878443 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.878530 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.878551 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878622 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:03:27.878591363 +0000 UTC m=+29.482930008 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878662 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878676 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878686 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878719 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:27.878711106 +0000 UTC m=+29.483049701 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878759 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878813 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:27.878795349 +0000 UTC m=+29.483133984 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878768 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878853 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878873 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:23 crc kubenswrapper[4735]: E1122 08:03:23.878923 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:27.878907322 +0000 UTC m=+29.483245967 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.882382 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.907771 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.922183 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.948406 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.973534 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:23 crc kubenswrapper[4735]: I1122 08:03:23.989484 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:23Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.008515 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.262634 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:24 crc kubenswrapper[4735]: E1122 08:03:24.262790 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.262854 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.262897 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:24 crc kubenswrapper[4735]: E1122 08:03:24.263005 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:24 crc kubenswrapper[4735]: E1122 08:03:24.263099 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.486092 4735 generic.go:334] "Generic (PLEG): container finished" podID="5003f416-41bd-48b2-8965-292d0f933500" containerID="de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887" exitCode=0 Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.486173 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerDied","Data":"de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887"} Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.488921 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307"} Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.488968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b"} Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.504287 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.530702 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.552738 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.577812 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.591670 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.623983 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.646624 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.658935 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.679156 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.694847 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.711662 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.725630 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.743779 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:24 crc kubenswrapper[4735]: I1122 08:03:24.760037 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:24Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.497495 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerStarted","Data":"cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76"} Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.507254 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f"} Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.507329 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770"} Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.507350 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6"} Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.527411 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.546962 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.561697 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.578071 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.591804 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.605733 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.618921 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.633612 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.669135 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.684185 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.704180 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.715784 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.726686 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.742083 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.812486 4735 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.814122 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.814158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.814167 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.814273 4735 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.821796 4735 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.821994 4735 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.823091 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.823141 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.823156 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.823176 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.823189 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:25Z","lastTransitionTime":"2025-11-22T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:25 crc kubenswrapper[4735]: E1122 08:03:25.910270 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.913288 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.913325 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.913333 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.913348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.913357 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:25Z","lastTransitionTime":"2025-11-22T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:25 crc kubenswrapper[4735]: E1122 08:03:25.923952 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.928774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.928804 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.928813 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.928826 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.928834 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:25Z","lastTransitionTime":"2025-11-22T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:25 crc kubenswrapper[4735]: E1122 08:03:25.940840 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:25Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.944331 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.944356 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.944363 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.944378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:25 crc kubenswrapper[4735]: I1122 08:03:25.944394 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:25Z","lastTransitionTime":"2025-11-22T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: E1122 08:03:26.005705 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.008632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.008653 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.008661 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.008673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.008681 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: E1122 08:03:26.042536 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: E1122 08:03:26.042856 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.055206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.055239 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.055247 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.055261 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.055270 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.158268 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.158304 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.158312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.158328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.158337 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.261603 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-l2rhr"] Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.262004 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.262426 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:26 crc kubenswrapper[4735]: E1122 08:03:26.262533 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.262597 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.262610 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:26 crc kubenswrapper[4735]: E1122 08:03:26.262709 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:26 crc kubenswrapper[4735]: E1122 08:03:26.262807 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.263173 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.263205 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.263217 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.263231 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.263245 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.265436 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.265715 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.265982 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.266072 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.287802 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.305558 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.322907 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.359838 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.365975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.366031 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.366049 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.366081 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.366110 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.395760 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.401501 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4ns7\" (UniqueName: \"kubernetes.io/projected/646e7dd4-85e4-448c-bdea-4e5870f5530f-kube-api-access-d4ns7\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.401550 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/646e7dd4-85e4-448c-bdea-4e5870f5530f-serviceca\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.401595 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/646e7dd4-85e4-448c-bdea-4e5870f5530f-host\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.412660 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.428027 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.440445 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.456128 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.469116 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.469176 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.469194 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.469217 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.469233 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.474078 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.488353 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.502420 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4ns7\" (UniqueName: \"kubernetes.io/projected/646e7dd4-85e4-448c-bdea-4e5870f5530f-kube-api-access-d4ns7\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.502492 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/646e7dd4-85e4-448c-bdea-4e5870f5530f-serviceca\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.502522 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/646e7dd4-85e4-448c-bdea-4e5870f5530f-host\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.502609 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/646e7dd4-85e4-448c-bdea-4e5870f5530f-host\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.504004 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/646e7dd4-85e4-448c-bdea-4e5870f5530f-serviceca\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.508856 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.513688 4735 generic.go:334] "Generic (PLEG): container finished" podID="5003f416-41bd-48b2-8965-292d0f933500" containerID="cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76" exitCode=0 Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.513739 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerDied","Data":"cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.518080 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.524749 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.526425 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4ns7\" (UniqueName: \"kubernetes.io/projected/646e7dd4-85e4-448c-bdea-4e5870f5530f-kube-api-access-d4ns7\") pod \"node-ca-l2rhr\" (UID: \"646e7dd4-85e4-448c-bdea-4e5870f5530f\") " pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.547450 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.559389 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.571934 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.572994 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.573035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.573048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.573067 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.573080 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.587450 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.598604 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.615262 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.629862 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.650840 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.665350 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.673990 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.680221 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.680260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.680271 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.680289 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.680306 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.692305 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.706757 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.725043 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.738531 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.752232 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.768616 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-l2rhr" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.773243 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.787067 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.787112 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.787125 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.787141 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.787152 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.797966 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:26Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.891015 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.891365 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.891378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.891394 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.891407 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.994984 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.995012 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.995020 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.995032 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:26 crc kubenswrapper[4735]: I1122 08:03:26.995042 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:26Z","lastTransitionTime":"2025-11-22T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.096865 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.096898 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.096909 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.096926 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.096936 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.199178 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.199222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.199235 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.199255 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.199267 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.303954 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.304010 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.304025 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.304101 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.304127 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.407172 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.407232 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.407249 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.407273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.407291 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.509653 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.509704 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.509715 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.509737 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.509750 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.524240 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-l2rhr" event={"ID":"646e7dd4-85e4-448c-bdea-4e5870f5530f","Type":"ContainerStarted","Data":"bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.524317 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-l2rhr" event={"ID":"646e7dd4-85e4-448c-bdea-4e5870f5530f","Type":"ContainerStarted","Data":"dcb41be7bacc0d5e542ce7de0b481cf62a2bf2c20b38daceb5a42023745dd6e2"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.529919 4735 generic.go:334] "Generic (PLEG): container finished" podID="5003f416-41bd-48b2-8965-292d0f933500" containerID="77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e" exitCode=0 Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.529964 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerDied","Data":"77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.542555 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.559166 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.576906 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.595927 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.611966 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.612005 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.612017 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.612038 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.612052 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.621376 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.634263 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.649657 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.661786 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.680301 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.697405 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.712800 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.714861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.714917 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.714936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.714972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.714994 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.747008 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.759348 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.772772 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.784499 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.801156 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.811444 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.818692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.818743 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.818760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.818784 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.818799 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.822190 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.822312 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.822380 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:35.822360654 +0000 UTC m=+37.426699259 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.829555 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.840191 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.854488 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.867252 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.879766 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.897290 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.902849 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.903544 4735 scope.go:117] "RemoveContainer" containerID="33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37" Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.903715 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.910870 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922655 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922688 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922699 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922714 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922727 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:27Z","lastTransitionTime":"2025-11-22T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922739 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922826 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922848 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.922873 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.922970 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923007 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:35.92299475 +0000 UTC m=+37.527333355 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923051 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:03:35.923045711 +0000 UTC m=+37.527384316 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923103 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923114 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923124 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923144 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:35.923138044 +0000 UTC m=+37.527476649 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923184 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923192 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923198 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:27 crc kubenswrapper[4735]: E1122 08:03:27.923217 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:35.923211236 +0000 UTC m=+37.527549841 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.930951 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.949316 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.960190 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.972241 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.982907 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:27 crc kubenswrapper[4735]: I1122 08:03:27.997719 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:27Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.025748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.025788 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.025797 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.025811 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.025819 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.129058 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.129113 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.129129 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.129153 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.129174 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.232959 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.233030 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.233043 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.233066 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.233078 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.262743 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.262798 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.262743 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:28 crc kubenswrapper[4735]: E1122 08:03:28.262981 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:28 crc kubenswrapper[4735]: E1122 08:03:28.263106 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:28 crc kubenswrapper[4735]: E1122 08:03:28.263177 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.335544 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.335576 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.335585 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.335597 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.335607 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.437936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.437968 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.437976 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.437990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.437999 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.538253 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerStarted","Data":"3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.539662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.539690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.539700 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.539716 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.539727 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.557748 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.574532 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.594432 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.608070 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.625237 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.641718 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.641764 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.641773 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.641788 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.641798 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.643753 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.656211 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.674596 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.687448 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.701352 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.717787 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.731717 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.743604 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.745086 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.745154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.745174 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.745200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.745219 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.758904 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.770312 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.848057 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.848089 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.848098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.848110 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.848120 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.952051 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.952431 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.952621 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.952802 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:28 crc kubenswrapper[4735]: I1122 08:03:28.952995 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:28Z","lastTransitionTime":"2025-11-22T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.056705 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.057000 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.057018 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.057044 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.057062 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.159218 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.159257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.159268 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.159283 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.159295 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.261559 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.261640 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.261658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.261683 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.261705 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.287089 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.300988 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.313320 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.346765 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.363629 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.364314 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.364491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.364622 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.364721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.364848 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.392501 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.404261 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.413412 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.425878 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.445543 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.461633 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.467052 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.467090 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.467100 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.467117 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.467128 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.475103 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.493141 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.507294 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.517319 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.542823 4735 generic.go:334] "Generic (PLEG): container finished" podID="5003f416-41bd-48b2-8965-292d0f933500" containerID="3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9" exitCode=0 Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.542870 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerDied","Data":"3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.550915 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.561405 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.569624 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.569893 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.569981 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.570056 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.570130 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.572987 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.595482 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.610868 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.624721 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.638144 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.652701 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.666241 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.672372 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.672414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.672425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.672439 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.672465 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.677950 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.693033 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.707943 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.723236 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.748041 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.761344 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.775222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.775263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.775274 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.775291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.775305 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.779172 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.879287 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.879362 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.879380 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.879414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.879435 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.982333 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.982498 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.982521 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.982545 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:29 crc kubenswrapper[4735]: I1122 08:03:29.982563 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:29Z","lastTransitionTime":"2025-11-22T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.084778 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.084817 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.084825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.084839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.084848 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.187409 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.187485 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.187509 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.187532 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.187545 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.262812 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.262892 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:30 crc kubenswrapper[4735]: E1122 08:03:30.262954 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.263033 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:30 crc kubenswrapper[4735]: E1122 08:03:30.263186 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:30 crc kubenswrapper[4735]: E1122 08:03:30.263334 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.289898 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.289939 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.289951 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.289972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.289985 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.392361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.392413 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.392426 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.392444 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.392479 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.495319 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.495374 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.495390 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.495414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.495430 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.557652 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerStarted","Data":"581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.579730 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.597632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.597706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.597730 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.597774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.597795 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.602955 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.623534 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.644060 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.663429 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.685920 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.701032 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.701071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.701080 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.701097 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.701106 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.701752 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.720878 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.742873 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.768555 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.785547 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.804060 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.804124 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.804141 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.804165 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.804181 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.810748 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.830030 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.846897 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.865136 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:30Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.906818 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.906867 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.906878 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.906896 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:30 crc kubenswrapper[4735]: I1122 08:03:30.906910 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:30Z","lastTransitionTime":"2025-11-22T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.008890 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.008952 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.008968 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.008990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.009010 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.112161 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.112214 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.112230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.112250 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.112267 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.214694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.215163 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.215183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.215206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.215221 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.317269 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.317320 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.317333 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.317356 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.317368 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.419766 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.419812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.419824 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.419841 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.419853 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.531281 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.531338 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.531355 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.531376 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.531391 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.566445 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.566906 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.566938 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.580402 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.595569 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.606065 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.627485 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.633978 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.634199 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.634257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.634315 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.634367 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.645923 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.657420 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.670015 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.679813 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.692077 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.700872 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.711338 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.721552 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.736369 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.736598 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.736710 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.736805 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.736883 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.736909 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.749239 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.779248 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.801682 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.816806 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.832211 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.839010 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.839043 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.839056 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.839077 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.839090 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.844507 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.856797 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.868166 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.879517 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.891740 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.910422 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.925518 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.939334 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.941344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.941395 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.941407 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.941427 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.941441 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:31Z","lastTransitionTime":"2025-11-22T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.955615 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.970409 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:31 crc kubenswrapper[4735]: I1122 08:03:31.991476 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:31Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.011929 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.021695 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.033849 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.043594 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.043881 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.044225 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.044408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.044557 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.054494 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.079639 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.101930 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.116636 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.134943 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.146367 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.147294 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.147344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.147358 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.147378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.147392 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.167079 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.177444 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.190609 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.206635 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.217555 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.235407 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.247573 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.249623 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.249667 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.249680 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.249697 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.249708 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.260216 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.262477 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.262514 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:32 crc kubenswrapper[4735]: E1122 08:03:32.262600 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:32 crc kubenswrapper[4735]: E1122 08:03:32.262670 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.262888 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:32 crc kubenswrapper[4735]: E1122 08:03:32.263167 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.281423 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.291667 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.352852 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.352894 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.352905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.352922 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.352933 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.456083 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.456139 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.456152 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.456200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.456216 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.558855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.558904 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.558912 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.558927 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.558935 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.577648 4735 generic.go:334] "Generic (PLEG): container finished" podID="5003f416-41bd-48b2-8965-292d0f933500" containerID="581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816" exitCode=0 Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.578611 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerDied","Data":"581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.607751 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.629690 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.648300 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.661716 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.661749 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.661757 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.661770 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.661779 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.667362 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.683152 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.698961 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.718487 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.740089 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.764781 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.764819 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.764828 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.764844 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.764855 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.765412 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.794124 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.813503 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.837196 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.853902 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.869181 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.869222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.869233 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.869252 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.869262 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.870474 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.890407 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:32Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.971998 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.972054 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.972066 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.972084 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:32 crc kubenswrapper[4735]: I1122 08:03:32.972098 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:32Z","lastTransitionTime":"2025-11-22T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.075265 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.075673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.075687 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.075703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.075714 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.179306 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.179357 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.179399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.179440 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.179553 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.282370 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.282445 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.282561 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.282600 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.282624 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.385722 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.385999 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.386019 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.386039 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.386057 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.488163 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.488205 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.488213 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.488228 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.488237 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.586204 4735 generic.go:334] "Generic (PLEG): container finished" podID="5003f416-41bd-48b2-8965-292d0f933500" containerID="3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d" exitCode=0 Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.586286 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerDied","Data":"3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.597535 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.597586 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.597598 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.597617 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.597630 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.605987 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.625258 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.643629 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.655307 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.670682 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.684919 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.699654 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.700526 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.700561 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.700574 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.700591 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.700605 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.713691 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.729312 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.748571 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.769541 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.785119 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.798051 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.803123 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.803155 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.803168 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.803185 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.803199 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.808059 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.820760 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:33Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.905877 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.905914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.905924 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.905938 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:33 crc kubenswrapper[4735]: I1122 08:03:33.905946 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:33Z","lastTransitionTime":"2025-11-22T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.008899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.008952 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.008969 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.008992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.009009 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.111479 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.111516 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.111525 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.111538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.111548 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.214447 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.214531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.214544 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.214563 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.214576 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.262914 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.262914 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:34 crc kubenswrapper[4735]: E1122 08:03:34.263127 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.262954 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:34 crc kubenswrapper[4735]: E1122 08:03:34.263223 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:34 crc kubenswrapper[4735]: E1122 08:03:34.263426 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.317426 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.317543 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.317571 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.317606 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.317632 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.421050 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.421100 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.421118 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.421137 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.421151 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.469806 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6"] Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.470570 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.474605 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.474609 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.492894 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.514847 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.524658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.524713 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.524731 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.524755 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.524773 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.536600 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.551913 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.573583 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.593184 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.608134 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2ac395c9-165f-473f-a4da-7655ae4e89a9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.608211 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghqm5\" (UniqueName: \"kubernetes.io/projected/2ac395c9-165f-473f-a4da-7655ae4e89a9-kube-api-access-ghqm5\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.608253 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2ac395c9-165f-473f-a4da-7655ae4e89a9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.608295 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2ac395c9-165f-473f-a4da-7655ae4e89a9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.611128 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.626847 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.627952 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.627984 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.628001 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.628018 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.628029 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.642734 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.673658 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.686728 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.708968 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2ac395c9-165f-473f-a4da-7655ae4e89a9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.709049 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghqm5\" (UniqueName: \"kubernetes.io/projected/2ac395c9-165f-473f-a4da-7655ae4e89a9-kube-api-access-ghqm5\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.709097 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2ac395c9-165f-473f-a4da-7655ae4e89a9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.709177 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2ac395c9-165f-473f-a4da-7655ae4e89a9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.710032 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2ac395c9-165f-473f-a4da-7655ae4e89a9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.710160 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2ac395c9-165f-473f-a4da-7655ae4e89a9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.717631 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2ac395c9-165f-473f-a4da-7655ae4e89a9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.719976 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.731059 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.731122 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.731137 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.731153 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.731165 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.738641 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghqm5\" (UniqueName: \"kubernetes.io/projected/2ac395c9-165f-473f-a4da-7655ae4e89a9-kube-api-access-ghqm5\") pod \"ovnkube-control-plane-749d76644c-cwlr6\" (UID: \"2ac395c9-165f-473f-a4da-7655ae4e89a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.738958 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.756871 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.769935 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.786597 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.793181 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:34Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:34 crc kubenswrapper[4735]: W1122 08:03:34.804219 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ac395c9_165f_473f_a4da_7655ae4e89a9.slice/crio-ce87f8b03c001794066e5853ed93220f855953a0ecd7ef3e5a449a1d3ba373e9 WatchSource:0}: Error finding container ce87f8b03c001794066e5853ed93220f855953a0ecd7ef3e5a449a1d3ba373e9: Status 404 returned error can't find the container with id ce87f8b03c001794066e5853ed93220f855953a0ecd7ef3e5a449a1d3ba373e9 Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.834694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.834767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.834788 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.834815 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.834832 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.937810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.937857 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.937873 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.937896 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:34 crc kubenswrapper[4735]: I1122 08:03:34.937914 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:34Z","lastTransitionTime":"2025-11-22T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.041288 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.041362 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.041389 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.041420 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.041444 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.145120 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.145183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.145202 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.145225 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.145242 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.218841 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-vvf57"] Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.219488 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.219569 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.237572 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.248643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.248704 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.248721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.248743 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.248761 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.253910 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.278298 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.295356 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.312424 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.314974 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqpfj\" (UniqueName: \"kubernetes.io/projected/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-kube-api-access-zqpfj\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.315077 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.331063 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.348425 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.353349 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.353397 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.353410 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.353429 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.353441 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.366179 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.384343 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.400695 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.416090 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqpfj\" (UniqueName: \"kubernetes.io/projected/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-kube-api-access-zqpfj\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.416159 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.416362 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.416598 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs podName:e9170f64-5b0d-4b2b-99ce-7ecb9f567620 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:35.916571412 +0000 UTC m=+37.520910057 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs") pod "network-metrics-daemon-vvf57" (UID: "e9170f64-5b0d-4b2b-99ce-7ecb9f567620") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.421200 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.434932 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.438092 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqpfj\" (UniqueName: \"kubernetes.io/projected/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-kube-api-access-zqpfj\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.455777 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.456396 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.456514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.456589 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.456671 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.456745 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.494607 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.510911 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.543149 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.559110 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.559159 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.559176 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.559199 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.559249 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.562020 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:35Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.598092 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" event={"ID":"5003f416-41bd-48b2-8965-292d0f933500","Type":"ContainerStarted","Data":"b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.599858 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" event={"ID":"2ac395c9-165f-473f-a4da-7655ae4e89a9","Type":"ContainerStarted","Data":"ce87f8b03c001794066e5853ed93220f855953a0ecd7ef3e5a449a1d3ba373e9"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.662950 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.663026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.663043 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.663072 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.663092 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.765769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.765810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.765821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.765836 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.765847 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.868066 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.868097 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.868105 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.868135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.868145 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.923010 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.923103 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.923143 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.923266 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.923339 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:51.92331648 +0000 UTC m=+53.527655135 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.923621 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.923781 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:51.923737741 +0000 UTC m=+53.528076426 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.924119 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:35 crc kubenswrapper[4735]: E1122 08:03:35.924194 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs podName:e9170f64-5b0d-4b2b-99ce-7ecb9f567620 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:36.924179333 +0000 UTC m=+38.528517978 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs") pod "network-metrics-daemon-vvf57" (UID: "e9170f64-5b0d-4b2b-99ce-7ecb9f567620") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.970997 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.971056 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.971073 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.971098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:35 crc kubenswrapper[4735]: I1122 08:03:35.971117 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:35Z","lastTransitionTime":"2025-11-22T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.023634 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.023889 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:03:52.023845964 +0000 UTC m=+53.628184629 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.024249 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.024304 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.024497 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.024541 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.024557 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.024646 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:52.024619294 +0000 UTC m=+53.628957899 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.024666 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.024706 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.024731 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.024955 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:52.024882681 +0000 UTC m=+53.629221426 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.074773 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.074822 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.074833 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.074855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.074870 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.176697 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.176751 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.176767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.176788 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.176803 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.263143 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.263228 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.263259 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.263349 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.263377 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.263544 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.279536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.279599 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.279612 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.279628 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.279649 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.382126 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.382174 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.382186 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.382204 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.382217 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.393766 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.393813 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.393823 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.393844 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.393854 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.413351 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.417597 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.417648 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.417661 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.417689 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.417702 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.438131 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.441944 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.441977 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.441988 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.442005 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.442016 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.457802 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.461579 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.461619 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.461632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.461649 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.461662 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.478124 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.481611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.481640 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.481651 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.481664 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.481674 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.502642 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.502755 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.504498 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.504540 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.504554 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.504572 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.504592 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.604523 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" event={"ID":"2ac395c9-165f-473f-a4da-7655ae4e89a9","Type":"ContainerStarted","Data":"69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.604601 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" event={"ID":"2ac395c9-165f-473f-a4da-7655ae4e89a9","Type":"ContainerStarted","Data":"027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.606086 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.606142 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.606158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.606177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.606191 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.621851 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.637052 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.650301 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.667713 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.682531 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.692875 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.705363 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.708816 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.708844 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.708852 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.708866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.708875 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.716234 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.728496 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.747410 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.762285 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.781504 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.794251 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.805764 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.810985 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.811059 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.811115 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.811140 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.811155 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.818740 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.833375 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.843817 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.856004 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.877870 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.892380 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.910268 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.913918 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.914044 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.914139 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.914204 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.914269 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:36Z","lastTransitionTime":"2025-11-22T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.922515 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.933980 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.934188 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:36 crc kubenswrapper[4735]: E1122 08:03:36.934256 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs podName:e9170f64-5b0d-4b2b-99ce-7ecb9f567620 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:38.934238307 +0000 UTC m=+40.538576912 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs") pod "network-metrics-daemon-vvf57" (UID: "e9170f64-5b0d-4b2b-99ce-7ecb9f567620") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.936904 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.952069 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.969387 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:36 crc kubenswrapper[4735]: I1122 08:03:36.986375 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:36Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.007362 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:37Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.017071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.017135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.017149 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.017171 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.017186 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.023316 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:37Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.038037 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:37Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.055624 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:37Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.066852 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:37Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.079868 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:37Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.093194 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:37Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.109367 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:37Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.120099 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.120145 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.120155 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.120170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.120183 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.222883 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.222950 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.222968 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.222989 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.223003 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.266734 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:37 crc kubenswrapper[4735]: E1122 08:03:37.266948 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.324931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.324989 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.325003 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.325026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.325041 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.427182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.427226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.427235 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.427254 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.427263 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.529590 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.529628 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.529638 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.529653 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.529664 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.631714 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.631758 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.631769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.631785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.631796 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.735423 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.735508 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.735525 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.735546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.735561 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.839153 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.839194 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.839207 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.839228 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.839250 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.942022 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.942263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.942344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.942434 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:37 crc kubenswrapper[4735]: I1122 08:03:37.942540 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:37Z","lastTransitionTime":"2025-11-22T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.045421 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.045495 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.045506 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.045520 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.045530 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.148405 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.148787 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.148905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.149002 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.149218 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.252160 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.252260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.252279 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.252312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.252331 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.262339 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.262339 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:38 crc kubenswrapper[4735]: E1122 08:03:38.262498 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:38 crc kubenswrapper[4735]: E1122 08:03:38.262559 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.262339 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:38 crc kubenswrapper[4735]: E1122 08:03:38.262614 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.358340 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.358381 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.358390 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.358405 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.358419 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.461225 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.461278 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.461290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.461309 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.461321 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.564391 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.564425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.564434 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.564448 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.564532 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.612971 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/0.log" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.615879 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0" exitCode=1 Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.615913 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.616561 4735 scope.go:117] "RemoveContainer" containerID="dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.633199 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.648401 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.665638 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.667333 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.667366 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.667378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.667395 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.667407 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.690541 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.703230 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.732308 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:38Z\\\",\\\"message\\\":\\\"2 08:03:37.913180 5978 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 08:03:37.913199 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 08:03:37.913206 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 08:03:37.913227 5978 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 08:03:37.913240 5978 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 08:03:37.913229 5978 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 08:03:37.913258 5978 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 08:03:37.913264 5978 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 08:03:37.913263 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 08:03:37.913270 5978 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:37.913285 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 08:03:37.913294 5978 factory.go:656] Stopping watch factory\\\\nI1122 08:03:37.913299 5978 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:37.913302 5978 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:37.913313 5978 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:37.913329 5978 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.743914 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.756310 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.767560 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.769530 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.769581 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.769594 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.769613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.769626 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.789249 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.803275 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.817872 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.835074 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.851097 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.868624 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.872216 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.872254 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.872266 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.872289 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.872300 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.896559 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.910965 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:38Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.954781 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:38 crc kubenswrapper[4735]: E1122 08:03:38.954955 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:38 crc kubenswrapper[4735]: E1122 08:03:38.955004 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs podName:e9170f64-5b0d-4b2b-99ce-7ecb9f567620 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:42.954989592 +0000 UTC m=+44.559328197 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs") pod "network-metrics-daemon-vvf57" (UID: "e9170f64-5b0d-4b2b-99ce-7ecb9f567620") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.973968 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.974048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.974068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.974095 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:38 crc kubenswrapper[4735]: I1122 08:03:38.974114 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:38Z","lastTransitionTime":"2025-11-22T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.076605 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.076648 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.076656 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.076671 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.076682 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.179780 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.179825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.179835 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.179853 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.179863 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.263050 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:39 crc kubenswrapper[4735]: E1122 08:03:39.263178 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.281414 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.281860 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.281905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.281914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.281933 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.281944 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.292597 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.310130 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.324261 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.340986 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.354996 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.367048 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.383875 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.384285 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.384305 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.384330 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.384350 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.387224 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.404668 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.419733 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.434240 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.446805 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.458575 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.469570 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.486737 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.486769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.486777 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.486790 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.486800 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.489510 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.502233 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.522755 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:38Z\\\",\\\"message\\\":\\\"2 08:03:37.913180 5978 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 08:03:37.913199 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 08:03:37.913206 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 08:03:37.913227 5978 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 08:03:37.913240 5978 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 08:03:37.913229 5978 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 08:03:37.913258 5978 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 08:03:37.913264 5978 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 08:03:37.913263 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 08:03:37.913270 5978 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:37.913285 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 08:03:37.913294 5978 factory.go:656] Stopping watch factory\\\\nI1122 08:03:37.913299 5978 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:37.913302 5978 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:37.913313 5978 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:37.913329 5978 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.588514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.588561 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.588572 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.588589 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.588601 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.621224 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/1.log" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.622161 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/0.log" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.626021 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8" exitCode=1 Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.626056 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.626087 4735 scope.go:117] "RemoveContainer" containerID="dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.626676 4735 scope.go:117] "RemoveContainer" containerID="0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8" Nov 22 08:03:39 crc kubenswrapper[4735]: E1122 08:03:39.626895 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.642986 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.657029 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.670256 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.681667 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.691066 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.691115 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.691129 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.691149 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.691165 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.703382 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:38Z\\\",\\\"message\\\":\\\"2 08:03:37.913180 5978 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 08:03:37.913199 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 08:03:37.913206 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 08:03:37.913227 5978 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 08:03:37.913240 5978 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 08:03:37.913229 5978 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 08:03:37.913258 5978 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 08:03:37.913264 5978 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 08:03:37.913263 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 08:03:37.913270 5978 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:37.913285 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 08:03:37.913294 5978 factory.go:656] Stopping watch factory\\\\nI1122 08:03:37.913299 5978 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:37.913302 5978 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:37.913313 5978 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:37.913329 5978 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:39Z\\\",\\\"message\\\":\\\".io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00748f997 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: authentication-operator,},ClusterIP:10.217.5.150,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.150],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1122 08:03:39.525603 6222 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525615 6222 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525622 6222 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dt2lg in node crc\\\\nI1122 08:03:39.525628 6222 obj_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.718470 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.740148 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.752241 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.761162 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.775040 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.785025 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.793745 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.793779 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.793787 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.793799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.793808 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.797270 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.807173 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.819065 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.831487 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.848585 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.859937 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:39Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.896126 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.896191 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.896210 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.896230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.896244 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.999706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.999774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.999793 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.999823 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:39 crc kubenswrapper[4735]: I1122 08:03:39.999848 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:39Z","lastTransitionTime":"2025-11-22T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.103012 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.103068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.103086 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.103110 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.103128 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.205760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.205806 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.206118 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.206137 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.206148 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.262511 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:40 crc kubenswrapper[4735]: E1122 08:03:40.262657 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.262792 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:40 crc kubenswrapper[4735]: E1122 08:03:40.262933 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.263252 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:40 crc kubenswrapper[4735]: E1122 08:03:40.263440 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.308593 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.308660 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.308681 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.308706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.308724 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.411413 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.411495 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.411509 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.411527 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.411542 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.514051 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.514166 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.514184 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.514209 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.514227 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.617148 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.617226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.617250 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.617280 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.617302 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.631640 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/1.log" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.719475 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.719514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.719530 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.719547 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.719556 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.821636 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.821735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.821775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.821807 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.821829 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.924587 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.924630 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.924645 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.924662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:40 crc kubenswrapper[4735]: I1122 08:03:40.924671 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:40Z","lastTransitionTime":"2025-11-22T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.027293 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.027336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.027348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.027368 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.027382 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.130687 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.130747 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.130767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.130793 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.130810 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.233720 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.233831 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.233856 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.233886 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.233910 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.262824 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:41 crc kubenswrapper[4735]: E1122 08:03:41.263246 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.336910 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.336990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.337020 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.337052 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.337077 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.439899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.439940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.439953 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.439970 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.439980 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.542658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.542694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.542708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.542723 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.542733 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.646156 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.646207 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.646218 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.646237 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.646250 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.749975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.750047 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.750069 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.750098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.750118 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.854261 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.854730 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.854980 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.855224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.855655 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.958803 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.958844 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.958854 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.958870 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:41 crc kubenswrapper[4735]: I1122 08:03:41.958878 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:41Z","lastTransitionTime":"2025-11-22T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.061924 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.061974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.061987 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.062009 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.062024 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.165979 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.166110 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.166135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.166166 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.166188 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.262522 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.262558 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.262682 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:42 crc kubenswrapper[4735]: E1122 08:03:42.262672 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:42 crc kubenswrapper[4735]: E1122 08:03:42.262801 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:42 crc kubenswrapper[4735]: E1122 08:03:42.263043 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.269339 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.269537 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.269562 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.269588 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.269615 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.372370 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.372429 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.372449 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.372526 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.372553 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.475169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.475232 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.475253 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.475278 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.475297 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.578538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.578605 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.578627 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.578657 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.578678 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.682329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.682450 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.682521 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.682595 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.682619 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.785968 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.786042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.786060 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.786089 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.786115 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.888972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.889058 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.889078 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.889104 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.889122 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.992926 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.992981 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.992995 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.993015 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:42 crc kubenswrapper[4735]: I1122 08:03:42.993029 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:42Z","lastTransitionTime":"2025-11-22T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.009990 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:43 crc kubenswrapper[4735]: E1122 08:03:43.010191 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:43 crc kubenswrapper[4735]: E1122 08:03:43.010263 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs podName:e9170f64-5b0d-4b2b-99ce-7ecb9f567620 nodeName:}" failed. No retries permitted until 2025-11-22 08:03:51.010242077 +0000 UTC m=+52.614580692 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs") pod "network-metrics-daemon-vvf57" (UID: "e9170f64-5b0d-4b2b-99ce-7ecb9f567620") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.094938 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.094986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.094999 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.095018 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.095030 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.197764 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.197820 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.197840 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.197862 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.197878 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.263133 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:43 crc kubenswrapper[4735]: E1122 08:03:43.263577 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.263836 4735 scope.go:117] "RemoveContainer" containerID="33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.300636 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.301337 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.301546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.301700 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.301861 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.404097 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.404124 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.404133 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.404147 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.404157 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.505924 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.505961 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.505969 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.505985 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.505995 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.609192 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.609248 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.609270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.609299 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.609318 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.713064 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.713123 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.713141 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.713169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.713184 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.815986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.816035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.816044 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.816061 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.816075 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.920137 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.920165 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.920174 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.920186 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:43 crc kubenswrapper[4735]: I1122 08:03:43.920193 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:43Z","lastTransitionTime":"2025-11-22T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.023148 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.023208 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.023228 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.023251 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.023267 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.131542 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.131619 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.131645 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.131679 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.131716 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.235629 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.236175 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.236432 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.236644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.236895 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.262379 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.262502 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:44 crc kubenswrapper[4735]: E1122 08:03:44.262605 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:44 crc kubenswrapper[4735]: E1122 08:03:44.262767 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.262919 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:44 crc kubenswrapper[4735]: E1122 08:03:44.263057 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.339693 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.339742 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.339758 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.339781 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.339798 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.443027 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.443329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.443492 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.443614 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.443710 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.545868 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.545920 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.545930 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.545947 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.545956 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.649053 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.649129 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.649148 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.649172 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.649194 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.651857 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.653965 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.654474 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.684627 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:38Z\\\",\\\"message\\\":\\\"2 08:03:37.913180 5978 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 08:03:37.913199 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 08:03:37.913206 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 08:03:37.913227 5978 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 08:03:37.913240 5978 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 08:03:37.913229 5978 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 08:03:37.913258 5978 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 08:03:37.913264 5978 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 08:03:37.913263 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 08:03:37.913270 5978 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:37.913285 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 08:03:37.913294 5978 factory.go:656] Stopping watch factory\\\\nI1122 08:03:37.913299 5978 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:37.913302 5978 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:37.913313 5978 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:37.913329 5978 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:39Z\\\",\\\"message\\\":\\\".io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00748f997 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: authentication-operator,},ClusterIP:10.217.5.150,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.150],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1122 08:03:39.525603 6222 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525615 6222 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525622 6222 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dt2lg in node crc\\\\nI1122 08:03:39.525628 6222 obj_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.700522 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.724330 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.739985 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.752403 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.752512 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.752533 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.752564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.752580 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.755836 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.771991 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.788747 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.812324 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.827773 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.842706 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.855051 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.855122 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.855137 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.855158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.855176 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.863401 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.879481 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.892599 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.910679 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.926614 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.945931 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.957682 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.957735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.957748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.957768 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.957780 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:44Z","lastTransitionTime":"2025-11-22T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:44 crc kubenswrapper[4735]: I1122 08:03:44.960910 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:44Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.060052 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.060348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.060507 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.060660 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.060795 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.163273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.163633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.163710 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.163791 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.163861 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.262782 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:45 crc kubenswrapper[4735]: E1122 08:03:45.263268 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.266540 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.266587 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.266598 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.266616 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.266628 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.368863 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.369422 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.369819 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.370121 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.370422 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.473806 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.473855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.473863 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.473879 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.473888 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.576857 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.577227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.577719 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.577888 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.578021 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.681726 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.681814 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.681839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.681875 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.681900 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.784751 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.784790 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.784799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.784814 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.784825 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.887331 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.887378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.887388 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.887404 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.887413 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.990768 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.990846 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.990865 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.990912 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:45 crc kubenswrapper[4735]: I1122 08:03:45.990942 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:45Z","lastTransitionTime":"2025-11-22T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.094329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.094385 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.094403 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.094425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.094443 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.197952 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.197997 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.198011 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.198035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.198055 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.263412 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.263447 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.263661 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.263724 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.263890 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.264003 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.301407 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.301508 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.301528 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.301552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.301574 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.404421 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.404503 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.404527 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.404554 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.404575 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.507409 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.507839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.508016 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.508224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.508373 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.611415 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.611859 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.612038 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.612184 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.612313 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.669774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.669861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.669889 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.669925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.669951 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.698105 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:46Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.703283 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.703342 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.703362 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.703388 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.703405 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.725027 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:46Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.730609 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.730896 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.731054 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.731236 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.731423 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.754100 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:46Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.760337 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.760402 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.760426 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.760455 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.760515 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.782912 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:46Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.790067 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.790138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.790158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.790187 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.790206 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.814130 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:46Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:46 crc kubenswrapper[4735]: E1122 08:03:46.814999 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.818111 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.818341 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.818538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.818698 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.818837 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.922711 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.922787 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.922805 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.922830 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:46 crc kubenswrapper[4735]: I1122 08:03:46.922849 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:46Z","lastTransitionTime":"2025-11-22T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.026377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.026425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.026436 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.026452 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.026488 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.129625 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.130034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.130124 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.130215 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.130282 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.232925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.232990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.233011 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.233041 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.233065 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.263285 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:47 crc kubenswrapper[4735]: E1122 08:03:47.263562 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.336224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.336512 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.341953 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.342595 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.342699 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.446706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.446977 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.447051 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.447111 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.447170 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.550229 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.550308 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.550328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.550359 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.550384 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.654441 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.654552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.654576 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.654607 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.654632 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.757637 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.757709 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.757735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.757765 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.757788 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.860942 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.861042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.861068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.861098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.861116 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.962875 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.962899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.962906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.962918 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:47 crc kubenswrapper[4735]: I1122 08:03:47.962927 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:47Z","lastTransitionTime":"2025-11-22T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.066204 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.066290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.066329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.066359 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.066376 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.164195 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.174400 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.174491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.174518 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.174549 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.174573 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.175427 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.187215 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.207285 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.227824 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.248500 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.263589 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.263706 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:48 crc kubenswrapper[4735]: E1122 08:03:48.263743 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:48 crc kubenswrapper[4735]: E1122 08:03:48.263855 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.264002 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:48 crc kubenswrapper[4735]: E1122 08:03:48.264153 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.268668 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.277182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.277235 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.277248 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.277266 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.277280 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.281749 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.298404 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.313647 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.330553 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.355389 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.369981 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.379899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.379959 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.379976 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.380003 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.380018 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.396183 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:38Z\\\",\\\"message\\\":\\\"2 08:03:37.913180 5978 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 08:03:37.913199 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 08:03:37.913206 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 08:03:37.913227 5978 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 08:03:37.913240 5978 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 08:03:37.913229 5978 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 08:03:37.913258 5978 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 08:03:37.913264 5978 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 08:03:37.913263 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 08:03:37.913270 5978 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:37.913285 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 08:03:37.913294 5978 factory.go:656] Stopping watch factory\\\\nI1122 08:03:37.913299 5978 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:37.913302 5978 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:37.913313 5978 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:37.913329 5978 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:39Z\\\",\\\"message\\\":\\\".io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00748f997 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: authentication-operator,},ClusterIP:10.217.5.150,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.150],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1122 08:03:39.525603 6222 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525615 6222 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525622 6222 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dt2lg in node crc\\\\nI1122 08:03:39.525628 6222 obj_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.409233 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.423576 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.439083 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.462135 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.482438 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:48Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.483845 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.483886 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.483897 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.483941 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.483956 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.586926 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.586986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.587006 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.587033 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.587055 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.690244 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.690295 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.690307 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.690322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.690331 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.793876 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.793927 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.793939 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.793957 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.793969 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.896549 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.896643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.896712 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.896745 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:48 crc kubenswrapper[4735]: I1122 08:03:48.896765 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:48Z","lastTransitionTime":"2025-11-22T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.006555 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.006616 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.006634 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.006659 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.006677 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.110241 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.110307 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.110331 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.110361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.110440 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.214583 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.214643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.214659 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.214683 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.214700 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.262676 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:49 crc kubenswrapper[4735]: E1122 08:03:49.263264 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.281249 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.303906 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.317087 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.317201 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.317226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.317256 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.317277 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.324306 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.343608 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.362069 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.377489 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.393023 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.406716 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.420272 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.420361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.420384 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.420420 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.420446 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.426929 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.449861 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.462516 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.473769 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.498149 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dffb3ac1e17f7d43746b936c977fe751c9097235ba6ee3699969f676c47fcef0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:38Z\\\",\\\"message\\\":\\\"2 08:03:37.913180 5978 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 08:03:37.913199 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 08:03:37.913206 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 08:03:37.913227 5978 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 08:03:37.913240 5978 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 08:03:37.913229 5978 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 08:03:37.913258 5978 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 08:03:37.913264 5978 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 08:03:37.913263 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 08:03:37.913270 5978 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:37.913285 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 08:03:37.913294 5978 factory.go:656] Stopping watch factory\\\\nI1122 08:03:37.913299 5978 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:37.913302 5978 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:37.913313 5978 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:37.913329 5978 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:39Z\\\",\\\"message\\\":\\\".io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00748f997 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: authentication-operator,},ClusterIP:10.217.5.150,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.150],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1122 08:03:39.525603 6222 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525615 6222 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525622 6222 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dt2lg in node crc\\\\nI1122 08:03:39.525628 6222 obj_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.509325 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.522728 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.525172 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.525224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.525238 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.525258 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.525270 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.533347 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.553393 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.564021 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:49Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.627484 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.627744 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.627806 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.627925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.627991 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.730806 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.731108 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.731178 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.731257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.731315 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.833894 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.834213 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.834301 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.834382 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.834450 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.937907 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.937972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.937994 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.938024 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:49 crc kubenswrapper[4735]: I1122 08:03:49.938045 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:49Z","lastTransitionTime":"2025-11-22T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.041662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.041711 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.041727 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.041747 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.041763 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.144492 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.144568 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.144592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.144620 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.144642 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.248974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.249031 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.249049 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.249072 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.249091 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.262675 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.262750 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:50 crc kubenswrapper[4735]: E1122 08:03:50.262826 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:50 crc kubenswrapper[4735]: E1122 08:03:50.262945 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.262750 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:50 crc kubenswrapper[4735]: E1122 08:03:50.263066 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.352632 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.352704 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.352721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.352747 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.352768 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.455668 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.455743 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.455768 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.455799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.455820 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.565641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.565695 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.565712 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.565739 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.565756 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.669211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.669302 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.669329 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.669365 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.669388 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.773691 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.773738 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.773749 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.773771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.773786 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.877913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.877986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.878011 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.878050 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.878079 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.981917 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.981987 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.982008 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.982037 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:50 crc kubenswrapper[4735]: I1122 08:03:50.982064 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:50Z","lastTransitionTime":"2025-11-22T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.018156 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:51 crc kubenswrapper[4735]: E1122 08:03:51.018408 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:51 crc kubenswrapper[4735]: E1122 08:03:51.018588 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs podName:e9170f64-5b0d-4b2b-99ce-7ecb9f567620 nodeName:}" failed. No retries permitted until 2025-11-22 08:04:07.018550539 +0000 UTC m=+68.622889214 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs") pod "network-metrics-daemon-vvf57" (UID: "e9170f64-5b0d-4b2b-99ce-7ecb9f567620") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.085868 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.086633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.086668 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.086701 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.086727 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.190216 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.190288 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.190311 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.190341 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.190366 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.262448 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:51 crc kubenswrapper[4735]: E1122 08:03:51.262724 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.268634 4735 scope.go:117] "RemoveContainer" containerID="0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.290369 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.293774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.293825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.293845 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.293874 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.293892 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.315509 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.341051 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.359630 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.381943 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.398130 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.398179 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.398195 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.398219 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.398234 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.400114 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.415286 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.429893 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.445578 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.460969 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.478262 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.498036 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.501369 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.501417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.501434 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.501488 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.501507 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.514275 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.538216 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.559017 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.576156 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.601889 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:39Z\\\",\\\"message\\\":\\\".io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00748f997 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: authentication-operator,},ClusterIP:10.217.5.150,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.150],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1122 08:03:39.525603 6222 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525615 6222 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525622 6222 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dt2lg in node crc\\\\nI1122 08:03:39.525628 6222 obj_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.606080 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.606174 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.606190 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.606210 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.606773 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.617569 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.682336 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/1.log" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.685449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.685895 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.702332 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.709993 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.710015 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.710023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.710035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.710044 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.716405 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.728639 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.738362 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.758400 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:39Z\\\",\\\"message\\\":\\\".io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00748f997 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: authentication-operator,},ClusterIP:10.217.5.150,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.150],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1122 08:03:39.525603 6222 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525615 6222 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525622 6222 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dt2lg in node crc\\\\nI1122 08:03:39.525628 6222 obj_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.774452 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.802840 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.812616 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.812658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.812669 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.812686 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.812702 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.820587 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.845710 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.861750 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.879977 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.895042 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.909713 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.914916 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.914962 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.914979 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.914999 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.915011 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:51Z","lastTransitionTime":"2025-11-22T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.925351 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.940412 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.947645 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.947689 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:51 crc kubenswrapper[4735]: E1122 08:03:51.947775 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:51 crc kubenswrapper[4735]: E1122 08:03:51.947794 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:51 crc kubenswrapper[4735]: E1122 08:03:51.947842 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:04:23.947822025 +0000 UTC m=+85.552160630 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:03:51 crc kubenswrapper[4735]: E1122 08:03:51.947915 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:04:23.947895907 +0000 UTC m=+85.552234512 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.956744 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.967946 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:51 crc kubenswrapper[4735]: I1122 08:03:51.981594 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:51Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.022193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.022226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.022237 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.022251 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.022261 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.048429 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.048923 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:04:24.048891923 +0000 UTC m=+85.653230548 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.049108 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.049173 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.049284 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.049310 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.049322 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.049375 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 08:04:24.049358625 +0000 UTC m=+85.653697230 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.049393 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.049418 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.049437 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.049630 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 08:04:24.04954786 +0000 UTC m=+85.653886615 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.152059 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.152123 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.152135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.152150 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.152166 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.255743 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.255799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.255813 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.255834 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.255850 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.263080 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.263083 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.263231 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.263282 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.263386 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.263505 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.358248 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.358339 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.358352 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.358370 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.358382 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.463130 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.463718 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.463732 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.463761 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.463774 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.566350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.566404 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.566414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.566431 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.566440 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.668968 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.669052 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.669070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.669096 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.669117 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.701121 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/2.log" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.702385 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/1.log" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.706851 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36" exitCode=1 Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.706908 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.706949 4735 scope.go:117] "RemoveContainer" containerID="0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.707719 4735 scope.go:117] "RemoveContainer" containerID="382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36" Nov 22 08:03:52 crc kubenswrapper[4735]: E1122 08:03:52.707945 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.731647 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.748055 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.766610 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.771865 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.771922 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.771941 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.771965 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.771985 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.782834 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.807499 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.829627 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.851885 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.868348 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.875088 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.875123 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.875139 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.875162 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.875179 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.888250 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.910167 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0504e99b6bb7028fcb557e7b4620af0001184980b3d342f258b09e49f4e221b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:39Z\\\",\\\"message\\\":\\\".io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00748f997 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: authentication-operator,},ClusterIP:10.217.5.150,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.150],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1122 08:03:39.525603 6222 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525615 6222 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-dt2lg\\\\nI1122 08:03:39.525622 6222 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-dt2lg in node crc\\\\nI1122 08:03:39.525628 6222 obj_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.958287 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.977312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.977346 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.977357 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.977373 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.977385 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:52Z","lastTransitionTime":"2025-11-22T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:52 crc kubenswrapper[4735]: I1122 08:03:52.992296 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:52Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.018132 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.033549 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.046489 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.058511 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.069601 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.080738 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.080815 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.080833 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.080857 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.080876 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.084211 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.184954 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.185017 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.185036 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.185064 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.185083 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.262520 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:53 crc kubenswrapper[4735]: E1122 08:03:53.262770 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.288159 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.288242 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.288263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.288291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.288311 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.391150 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.391193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.391209 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.391231 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.391247 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.494257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.494314 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.494336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.494365 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.494391 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.597722 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.597784 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.597807 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.597835 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.597856 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.701741 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.701825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.701852 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.701885 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.701911 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.716323 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/2.log" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.723169 4735 scope.go:117] "RemoveContainer" containerID="382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36" Nov 22 08:03:53 crc kubenswrapper[4735]: E1122 08:03:53.723448 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.745989 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.764121 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.789739 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.804775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.804832 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.804850 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.804878 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.804898 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.809382 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.835334 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.857947 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.879401 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.894625 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.907681 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.907731 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.907744 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.907764 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.907778 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:53Z","lastTransitionTime":"2025-11-22T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.911775 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.926752 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.952073 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.967208 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.981040 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:53 crc kubenswrapper[4735]: I1122 08:03:53.999318 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:53Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.010746 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.010783 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.010795 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.010810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.010819 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.013020 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:54Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.022137 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:54Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.039832 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:54Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.052675 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:54Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.114225 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.114315 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.114342 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.114373 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.114397 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.217538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.217588 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.217599 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.217616 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.217628 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.262802 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.263146 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:54 crc kubenswrapper[4735]: E1122 08:03:54.263358 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.263417 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:54 crc kubenswrapper[4735]: E1122 08:03:54.263501 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:54 crc kubenswrapper[4735]: E1122 08:03:54.263615 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.321596 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.321640 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.321649 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.321664 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.321674 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.424116 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.424200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.424215 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.424236 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.424249 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.526315 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.526356 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.526365 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.526379 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.526390 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.629271 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.629330 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.629348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.629373 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.629391 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.732180 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.732255 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.732275 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.732301 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.732322 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.835987 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.836336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.836512 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.836621 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.836706 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.940642 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.940732 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.940752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.940779 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:54 crc kubenswrapper[4735]: I1122 08:03:54.940801 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:54Z","lastTransitionTime":"2025-11-22T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.043774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.043816 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.043825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.043840 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.043850 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.147189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.147236 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.147247 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.147273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.147290 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.250259 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.250443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.250588 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.250652 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.250680 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.262970 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:55 crc kubenswrapper[4735]: E1122 08:03:55.263159 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.354118 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.354165 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.354176 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.354199 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.354210 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.457573 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.457631 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.457647 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.457673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.457691 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.560919 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.560992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.561017 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.561051 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.561068 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.664509 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.664613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.664633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.664668 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.664690 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.766800 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.766846 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.766855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.766873 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.766883 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.870771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.870918 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.870934 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.870960 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.870974 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.973830 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.973877 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.973889 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.973906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:55 crc kubenswrapper[4735]: I1122 08:03:55.973919 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:55Z","lastTransitionTime":"2025-11-22T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.076944 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.076989 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.077029 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.077051 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.077063 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.180318 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.180394 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.180417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.180445 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.180533 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.263158 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.263181 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.263168 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:56 crc kubenswrapper[4735]: E1122 08:03:56.263327 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:56 crc kubenswrapper[4735]: E1122 08:03:56.263493 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:56 crc kubenswrapper[4735]: E1122 08:03:56.263630 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.283609 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.283664 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.283680 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.283702 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.283720 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.385833 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.385874 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.385884 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.385899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.385910 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.489440 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.489610 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.489638 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.489666 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.489683 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.592636 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.592675 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.592684 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.592701 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.592716 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.695291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.695327 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.695340 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.695358 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.695373 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.798498 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.798550 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.798566 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.798590 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.798607 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.856783 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.873702 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:56Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.890433 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:56Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.901265 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.901331 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.901353 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.901377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.901395 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:56Z","lastTransitionTime":"2025-11-22T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.918001 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:56Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.934988 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:56Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.957667 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:56Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.973240 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:56Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:56 crc kubenswrapper[4735]: I1122 08:03:56.987922 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:56Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.004031 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.004871 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.004932 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.004947 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.004968 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.004981 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.019659 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.031361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.031435 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.031446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.031501 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.031513 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.034189 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: E1122 08:03:57.045244 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.048700 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.048786 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.048827 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.048849 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.048861 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.049565 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.063347 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: E1122 08:03:57.067609 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.071395 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.071435 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.071489 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.071516 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.071527 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.080486 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: E1122 08:03:57.085559 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.089273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.089305 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.089315 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.089336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.089348 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.099817 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: E1122 08:03:57.108057 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.114702 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.114744 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.114756 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.114778 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.114789 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.118135 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.132476 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: E1122 08:03:57.132921 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: E1122 08:03:57.133093 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.134526 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.134568 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.134578 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.134593 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.134603 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.160510 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.173644 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:57Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.236578 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.236619 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.236630 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.236646 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.236656 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.262479 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:57 crc kubenswrapper[4735]: E1122 08:03:57.262592 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.339873 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.339920 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.339936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.339960 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.339977 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.442861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.442922 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.442938 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.442964 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.442983 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.546238 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.546296 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.546315 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.546339 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.546359 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.649348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.649427 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.649452 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.649536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.649562 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.752234 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.752304 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.752325 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.752351 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.752372 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.854916 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.854958 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.854969 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.854983 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.854993 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.957721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.957859 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.957879 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.957902 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:57 crc kubenswrapper[4735]: I1122 08:03:57.957919 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:57Z","lastTransitionTime":"2025-11-22T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.060772 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.060814 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.060824 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.060837 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.060847 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.163203 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.163234 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.163242 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.163256 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.163264 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.262295 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.262365 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.262429 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:03:58 crc kubenswrapper[4735]: E1122 08:03:58.262608 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:03:58 crc kubenswrapper[4735]: E1122 08:03:58.262723 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:03:58 crc kubenswrapper[4735]: E1122 08:03:58.262829 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.265493 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.265525 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.265537 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.265554 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.265565 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.367806 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.367861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.367874 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.367892 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.367904 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.470698 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.470769 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.470801 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.470832 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.470853 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.573869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.573955 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.573977 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.574002 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.574021 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.677782 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.677861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.677886 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.677914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.677934 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.780532 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.780593 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.780609 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.780633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.780655 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.883513 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.883572 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.883589 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.883613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.883630 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.986127 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.986181 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.986190 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.986205 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:58 crc kubenswrapper[4735]: I1122 08:03:58.986215 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:58Z","lastTransitionTime":"2025-11-22T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.088678 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.088767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.088796 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.088827 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.088851 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.196988 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.197023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.197031 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.197044 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.197052 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.262498 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:03:59 crc kubenswrapper[4735]: E1122 08:03:59.262697 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.288423 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.299839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.299899 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.299914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.299936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.299954 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.308355 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.328627 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.363806 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.402399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.402504 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.402521 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.402544 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.402558 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.404449 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.426700 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.450891 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.469243 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.479154 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.489585 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.502358 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.505186 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.505251 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.505273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.505305 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.505327 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.517969 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.537171 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.550129 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.566942 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.581680 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.597957 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.607382 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.607419 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.607431 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.607449 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.607481 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.614528 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:03:59Z is after 2025-08-24T17:21:41Z" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.709709 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.709752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.709764 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.709780 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.709792 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.812852 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.812904 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.812915 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.812931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.812942 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.916863 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.916925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.916943 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.916966 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:03:59 crc kubenswrapper[4735]: I1122 08:03:59.916984 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:03:59Z","lastTransitionTime":"2025-11-22T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.019876 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.019931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.019947 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.019970 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.019987 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.122713 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.122752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.122765 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.122783 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.122797 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.225443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.225492 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.225500 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.225512 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.225520 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.262489 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.262584 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:00 crc kubenswrapper[4735]: E1122 08:04:00.262632 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.262501 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:00 crc kubenswrapper[4735]: E1122 08:04:00.262714 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:00 crc kubenswrapper[4735]: E1122 08:04:00.262805 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.328156 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.328206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.328218 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.328237 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.328253 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.431060 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.431106 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.431117 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.431136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.431148 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.533856 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.533936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.533954 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.533981 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.533998 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.637510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.637591 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.637615 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.637649 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.637666 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.740374 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.740438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.740488 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.740514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.740531 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.845484 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.845587 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.845612 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.845647 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.845682 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.947820 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.947888 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.947907 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.947931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:00 crc kubenswrapper[4735]: I1122 08:04:00.947948 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:00Z","lastTransitionTime":"2025-11-22T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.050328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.050399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.050420 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.050447 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.050495 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.153288 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.153367 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.153386 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.153414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.153435 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.256657 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.256729 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.256745 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.256775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.256787 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.263375 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:01 crc kubenswrapper[4735]: E1122 08:04:01.263602 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.359270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.359375 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.359397 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.359425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.359443 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.462314 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.462377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.462390 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.462408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.462422 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.565896 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.565966 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.565977 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.565998 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.566010 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.668810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.668885 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.668905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.668933 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.668954 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.771937 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.771993 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.772004 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.772019 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.772029 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.875624 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.875703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.875736 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.875767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.875789 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.979239 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.979322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.979342 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.979367 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:01 crc kubenswrapper[4735]: I1122 08:04:01.979384 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:01Z","lastTransitionTime":"2025-11-22T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.082845 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.082913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.082933 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.082958 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.082976 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.184913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.184974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.184991 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.185017 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.185035 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.263212 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.263456 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.263568 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:02 crc kubenswrapper[4735]: E1122 08:04:02.263615 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:02 crc kubenswrapper[4735]: E1122 08:04:02.263535 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:02 crc kubenswrapper[4735]: E1122 08:04:02.263953 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.287679 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.287750 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.287775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.287805 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.287829 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.390316 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.390355 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.390364 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.390376 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.390384 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.493577 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.493639 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.493666 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.493699 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.493717 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.595756 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.595818 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.595836 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.595859 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.595875 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.700517 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.700563 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.700574 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.700592 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.700605 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.802612 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.802660 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.802672 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.802690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.802701 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.905155 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.905188 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.905197 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.905212 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:02 crc kubenswrapper[4735]: I1122 08:04:02.905221 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:02Z","lastTransitionTime":"2025-11-22T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.010411 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.010490 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.010509 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.010532 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.010548 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.112793 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.112860 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.112878 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.112903 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.112921 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.215084 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.215121 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.215133 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.215148 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.215160 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.269165 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:03 crc kubenswrapper[4735]: E1122 08:04:03.269321 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.317661 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.317734 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.317757 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.317787 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.317806 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.420882 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.420924 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.420940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.420962 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.420978 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.524943 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.524984 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.524997 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.525015 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.525026 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.627227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.627256 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.627264 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.627278 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.627286 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.730986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.731034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.731050 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.731071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.731087 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.833532 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.833575 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.833589 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.833611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.833628 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.937263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.937342 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.937360 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.937386 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:03 crc kubenswrapper[4735]: I1122 08:04:03.937430 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:03Z","lastTransitionTime":"2025-11-22T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.042028 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.042071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.042080 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.042093 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.042102 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.145306 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.145360 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.145373 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.145394 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.145408 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.248371 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.248422 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.248434 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.248515 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.248530 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.262436 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.262521 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.262498 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:04 crc kubenswrapper[4735]: E1122 08:04:04.262697 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:04 crc kubenswrapper[4735]: E1122 08:04:04.262863 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:04 crc kubenswrapper[4735]: E1122 08:04:04.263036 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.351150 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.351212 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.351272 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.351301 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.351315 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.454673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.454731 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.454748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.454771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.454790 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.557007 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.557052 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.557064 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.557082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.557103 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.659689 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.659734 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.659752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.659772 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.659786 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.762057 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.762091 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.762101 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.762117 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.762127 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.865689 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.865736 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.865747 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.865765 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.865777 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.978842 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.978881 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.978891 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.978907 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:04 crc kubenswrapper[4735]: I1122 08:04:04.978917 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:04Z","lastTransitionTime":"2025-11-22T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.081951 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.081992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.082004 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.082023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.082039 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.184749 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.184808 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.184829 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.184854 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.184871 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.263144 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:05 crc kubenswrapper[4735]: E1122 08:04:05.263319 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.266134 4735 scope.go:117] "RemoveContainer" containerID="382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36" Nov 22 08:04:05 crc kubenswrapper[4735]: E1122 08:04:05.266830 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.287731 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.287788 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.287808 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.287834 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.287852 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.391771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.391810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.391819 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.391835 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.391844 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.494877 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.494914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.494923 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.494941 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.494951 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.600976 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.601010 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.601024 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.601039 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.601048 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.703334 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.703394 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.703411 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.703434 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.703451 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.805445 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.805487 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.805495 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.805508 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.805517 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.908068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.908132 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.908150 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.908178 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:05 crc kubenswrapper[4735]: I1122 08:04:05.908194 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:05Z","lastTransitionTime":"2025-11-22T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.010765 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.010821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.010838 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.010862 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.010879 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.114298 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.114369 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.114396 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.114426 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.114449 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.217227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.217292 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.217317 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.217350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.217372 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.262488 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.262604 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.263020 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:06 crc kubenswrapper[4735]: E1122 08:04:06.263256 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:06 crc kubenswrapper[4735]: E1122 08:04:06.263383 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:06 crc kubenswrapper[4735]: E1122 08:04:06.263446 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.320711 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.321042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.321185 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.321319 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.321433 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.424756 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.425025 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.425145 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.425290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.425406 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.527697 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.527768 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.527790 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.527820 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.527842 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.630739 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.630766 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.630774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.630786 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.630794 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.733161 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.733195 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.733203 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.733218 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.733231 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.834974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.835013 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.835026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.835044 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.835055 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.937140 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.937169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.937177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.937192 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:06 crc kubenswrapper[4735]: I1122 08:04:06.937202 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:06Z","lastTransitionTime":"2025-11-22T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.040240 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.040338 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.040365 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.040400 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.040429 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.117607 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.117752 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.117816 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs podName:e9170f64-5b0d-4b2b-99ce-7ecb9f567620 nodeName:}" failed. No retries permitted until 2025-11-22 08:04:39.11779793 +0000 UTC m=+100.722136595 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs") pod "network-metrics-daemon-vvf57" (UID: "e9170f64-5b0d-4b2b-99ce-7ecb9f567620") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.142753 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.142795 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.142806 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.142825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.142837 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.245348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.245681 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.245777 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.245890 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.245971 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.262983 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.263169 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.348825 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.348870 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.348882 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.348898 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.348909 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.451101 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.451177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.451199 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.451228 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.451251 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.524386 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.524431 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.524443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.524476 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.524489 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.547428 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:07Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.552082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.552131 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.552151 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.552170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.552185 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.571012 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:07Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.579846 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.579921 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.579941 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.579976 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.579994 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.595711 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:07Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.600672 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.600714 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.600731 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.600752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.600768 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.620753 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:07Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.625315 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.625443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.625531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.625602 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.625664 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.644786 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:07Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:07 crc kubenswrapper[4735]: E1122 08:04:07.645485 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.647399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.647514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.647546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.647579 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.647604 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.750679 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.750710 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.750718 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.750731 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.750740 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.852940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.853011 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.853023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.853037 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.853048 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.954996 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.955074 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.955098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.955136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:07 crc kubenswrapper[4735]: I1122 08:04:07.955158 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:07Z","lastTransitionTime":"2025-11-22T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.057882 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.057965 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.057999 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.058028 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.058049 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.164255 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.164348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.164365 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.164387 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.164404 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.262725 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.262769 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.262888 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:08 crc kubenswrapper[4735]: E1122 08:04:08.263075 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:08 crc kubenswrapper[4735]: E1122 08:04:08.263251 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:08 crc kubenswrapper[4735]: E1122 08:04:08.263376 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.267826 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.267890 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.267906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.268318 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.268373 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.371862 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.371901 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.371909 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.371923 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.371932 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.474996 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.475069 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.475091 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.475119 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.475173 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.577975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.578034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.578048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.578065 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.578077 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.681390 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.681443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.681455 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.681492 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.681508 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.784079 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.784140 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.784157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.784180 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.784198 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.886402 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.886437 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.886446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.886505 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.886515 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.990168 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.990236 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.990259 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.990287 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:08 crc kubenswrapper[4735]: I1122 08:04:08.990350 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:08Z","lastTransitionTime":"2025-11-22T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.092908 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.092979 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.093006 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.093034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.093056 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.201772 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.201821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.201834 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.201851 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.201860 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.262908 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:09 crc kubenswrapper[4735]: E1122 08:04:09.263179 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.281915 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.298529 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.305324 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.305408 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.305434 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.305499 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.305526 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.316574 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.335835 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.360238 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.373606 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.390424 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.402853 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.407579 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.407611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.407619 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.407634 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.407646 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.423045 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.447523 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.465720 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.481209 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.511159 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.511207 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.511224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.511245 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.511260 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.511479 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.521293 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.532188 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.540344 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.554765 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.563403 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:09Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.613411 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.613442 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.613449 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.613485 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.613493 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.715854 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.715889 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.715898 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.715911 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.715920 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.818452 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.818513 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.818524 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.818539 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.818549 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.920763 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.920847 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.920871 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.920890 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:09 crc kubenswrapper[4735]: I1122 08:04:09.920901 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:09Z","lastTransitionTime":"2025-11-22T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.023114 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.023152 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.023164 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.023180 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.023192 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.125424 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.125524 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.125536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.125555 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.125567 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.228098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.228134 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.228142 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.228154 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.228163 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.262623 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.262678 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:10 crc kubenswrapper[4735]: E1122 08:04:10.262744 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.262685 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:10 crc kubenswrapper[4735]: E1122 08:04:10.262830 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:10 crc kubenswrapper[4735]: E1122 08:04:10.262938 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.330699 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.330742 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.330755 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.330771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.330785 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.433929 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.434207 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.434269 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.434340 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.434402 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.536276 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.536345 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.536367 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.536398 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.536420 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.638886 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.638920 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.638931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.638949 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.638962 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.741929 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.742024 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.742035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.742050 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.742059 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.844919 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.845273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.845512 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.845717 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.845886 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.948188 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.948420 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.948503 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.948598 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:10 crc kubenswrapper[4735]: I1122 08:04:10.948674 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:10Z","lastTransitionTime":"2025-11-22T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.051916 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.052376 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.052484 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.052593 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.052669 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.155576 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.155611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.155622 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.155638 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.155649 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.257611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.257651 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.257660 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.257673 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.257683 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.262466 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:11 crc kubenswrapper[4735]: E1122 08:04:11.262587 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.359378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.359421 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.359433 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.359451 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.359473 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.461195 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.461422 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.461550 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.461619 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.461680 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.564742 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.564787 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.564797 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.564815 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.564827 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.667287 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.667815 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.667906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.667992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.668072 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.770514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.770587 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.770601 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.770622 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.770636 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.872472 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.872511 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.872520 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.872534 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.872542 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.976013 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.976495 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.976590 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.976689 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:11 crc kubenswrapper[4735]: I1122 08:04:11.976787 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:11Z","lastTransitionTime":"2025-11-22T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.079688 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.079768 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.079804 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.079831 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.079847 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.182108 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.182153 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.182162 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.182177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.182189 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.262507 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.263354 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:12 crc kubenswrapper[4735]: E1122 08:04:12.263484 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.263544 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:12 crc kubenswrapper[4735]: E1122 08:04:12.263736 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:12 crc kubenswrapper[4735]: E1122 08:04:12.263790 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.285192 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.285259 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.285275 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.285298 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.285316 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.388031 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.388064 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.388094 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.388128 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.388138 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.491355 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.491416 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.491433 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.491486 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.491507 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.594083 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.594123 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.594135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.594153 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.594166 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.697191 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.697270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.697282 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.697295 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.697305 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.780908 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/0.log" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.780965 4735 generic.go:334] "Generic (PLEG): container finished" podID="61dbab7f-c903-4b55-9d90-aacfd12f7f0f" containerID="04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5" exitCode=1 Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.781013 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-85z2h" event={"ID":"61dbab7f-c903-4b55-9d90-aacfd12f7f0f","Type":"ContainerDied","Data":"04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.781573 4735 scope.go:117] "RemoveContainer" containerID="04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.799269 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.799324 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.799334 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.799352 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.799364 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.803578 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.819164 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.832122 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.848830 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.865257 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.880140 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.897964 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.905839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.905906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.905925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.905950 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.905968 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:12Z","lastTransitionTime":"2025-11-22T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.912561 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.929718 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.941834 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.957004 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.970290 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:12 crc kubenswrapper[4735]: I1122 08:04:12.981661 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"2025-11-22T08:03:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d\\\\n2025-11-22T08:03:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d to /host/opt/cni/bin/\\\\n2025-11-22T08:03:27Z [verbose] multus-daemon started\\\\n2025-11-22T08:03:27Z [verbose] Readiness Indicator file check\\\\n2025-11-22T08:04:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.000723 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:12Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.008149 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.008179 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.008187 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.008201 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.008214 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.013085 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.033772 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.047168 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.058908 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.111090 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.111155 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.111166 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.111181 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.111189 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.213992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.214042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.214054 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.214072 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.214083 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.262630 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:13 crc kubenswrapper[4735]: E1122 08:04:13.263047 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.316606 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.316663 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.316674 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.316690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.316702 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.418936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.419256 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.419357 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.419473 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.419566 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.522267 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.522317 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.522331 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.522352 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.522365 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.625136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.625378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.625497 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.625574 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.625639 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.727699 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.727778 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.727792 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.727847 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.727867 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.787160 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/0.log" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.788067 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-85z2h" event={"ID":"61dbab7f-c903-4b55-9d90-aacfd12f7f0f","Type":"ContainerStarted","Data":"a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.807047 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.819111 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.830269 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.830317 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.830328 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.830350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.830365 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.834841 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.853581 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.870900 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.885327 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.900596 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.921686 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"2025-11-22T08:03:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d\\\\n2025-11-22T08:03:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d to /host/opt/cni/bin/\\\\n2025-11-22T08:03:27Z [verbose] multus-daemon started\\\\n2025-11-22T08:03:27Z [verbose] Readiness Indicator file check\\\\n2025-11-22T08:04:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:04:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.933776 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.933813 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.933822 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.933837 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.933846 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:13Z","lastTransitionTime":"2025-11-22T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.942793 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.961007 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:13 crc kubenswrapper[4735]: I1122 08:04:13.976980 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:13Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.007203 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:14Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.022409 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:14Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.035923 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.035967 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.035987 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.036010 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.036027 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.046272 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:14Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.058014 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:14Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.074531 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:14Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.085816 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:14Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.099835 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:14Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.138797 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.139049 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.139131 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.139194 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.139254 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.242014 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.242067 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.242084 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.242108 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.242125 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.262634 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.262677 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:14 crc kubenswrapper[4735]: E1122 08:04:14.262790 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:14 crc kubenswrapper[4735]: E1122 08:04:14.263000 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.263267 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:14 crc kubenswrapper[4735]: E1122 08:04:14.263437 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.345157 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.345219 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.345235 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.345262 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.345282 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.448005 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.448246 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.448324 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.448424 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.448507 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.552200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.552247 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.552260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.552278 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.552294 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.655089 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.655542 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.655704 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.655984 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.656152 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.759058 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.759152 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.759169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.759222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.759243 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.862180 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.862262 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.862286 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.862317 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.862339 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.965635 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.965697 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.965713 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.965737 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:14 crc kubenswrapper[4735]: I1122 08:04:14.965754 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:14Z","lastTransitionTime":"2025-11-22T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.068938 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.068981 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.068990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.069007 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.069021 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.171979 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.172384 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.172644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.172857 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.173067 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.263286 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:15 crc kubenswrapper[4735]: E1122 08:04:15.263549 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.275236 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.275303 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.275327 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.275358 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.275383 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.378336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.378411 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.378435 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.378502 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.378522 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.480861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.481179 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.481287 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.481382 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.481514 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.583985 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.584042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.584056 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.584076 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.584088 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.688707 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.689644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.689670 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.689690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.689702 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.793044 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.793144 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.793169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.793197 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.793215 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.896943 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.897637 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.897678 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.897705 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:15 crc kubenswrapper[4735]: I1122 08:04:15.897721 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:15Z","lastTransitionTime":"2025-11-22T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.001226 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.001268 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.001280 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.001297 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.001308 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.104357 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.104401 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.104411 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.104429 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.104440 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.207895 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.208000 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.208016 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.208319 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.208357 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.262787 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.262804 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:16 crc kubenswrapper[4735]: E1122 08:04:16.262994 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:16 crc kubenswrapper[4735]: E1122 08:04:16.263169 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.263251 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:16 crc kubenswrapper[4735]: E1122 08:04:16.263559 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.311413 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.311491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.311504 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.311524 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.311536 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.415166 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.415247 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.415271 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.415303 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.415324 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.519556 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.519931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.520069 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.520193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.520291 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.623505 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.623571 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.623590 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.623613 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.623629 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.726987 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.727035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.727048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.727066 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.727078 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.830158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.830230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.830248 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.830276 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.830291 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.933639 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.933694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.933707 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.933733 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:16 crc kubenswrapper[4735]: I1122 08:04:16.933766 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:16Z","lastTransitionTime":"2025-11-22T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.036107 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.036180 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.036204 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.036234 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.036255 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.139616 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.139672 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.139684 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.139709 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.139722 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.243006 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.243086 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.243105 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.243129 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.243144 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.263435 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:17 crc kubenswrapper[4735]: E1122 08:04:17.263927 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.264369 4735 scope.go:117] "RemoveContainer" containerID="382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.345169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.345565 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.345804 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.345978 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.346123 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.449275 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.449312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.449327 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.449348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.449364 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.552646 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.552715 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.552738 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.552767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.552794 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.654925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.654973 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.654983 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.655001 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.655013 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.756752 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.756778 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.756785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.756798 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.756807 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.804422 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/2.log" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.806862 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.808045 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.820428 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.834180 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.847297 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"2025-11-22T08:03:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d\\\\n2025-11-22T08:03:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d to /host/opt/cni/bin/\\\\n2025-11-22T08:03:27Z [verbose] multus-daemon started\\\\n2025-11-22T08:03:27Z [verbose] Readiness Indicator file check\\\\n2025-11-22T08:04:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:04:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.859263 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.859304 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.859316 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.859332 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.859343 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.867873 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.881355 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.891946 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.910671 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:04:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.927976 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.943139 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.955708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.955761 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.955771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.955793 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.955817 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.956030 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: E1122 08:04:17.970506 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.972401 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.975277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.975338 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.975350 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.975387 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.975402 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.985404 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: E1122 08:04:17.989944 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.993208 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.993247 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.993258 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.993277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.993290 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:17Z","lastTransitionTime":"2025-11-22T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:17 crc kubenswrapper[4735]: I1122 08:04:17.998591 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:17Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: E1122 08:04:18.010702 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.012399 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.014601 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.014644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.014656 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.014676 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.014688 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: E1122 08:04:18.026286 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.026588 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.029177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.029198 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.029208 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.029223 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.029236 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.041150 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: E1122 08:04:18.043951 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: E1122 08:04:18.044110 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.045289 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.045316 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.045325 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.045339 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.045348 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.055591 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.070911 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.147869 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.147897 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.147905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.147917 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.147925 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.251068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.251102 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.251114 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.251129 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.251140 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.262523 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.262566 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.262625 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:18 crc kubenswrapper[4735]: E1122 08:04:18.262634 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:18 crc kubenswrapper[4735]: E1122 08:04:18.262739 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:18 crc kubenswrapper[4735]: E1122 08:04:18.262801 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.353913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.353978 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.354000 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.354029 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.354050 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.456665 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.456712 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.456723 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.456741 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.456754 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.559624 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.559685 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.559695 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.559711 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.559722 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.663436 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.663519 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.663530 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.663548 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.663561 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.766536 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.766829 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.767034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.767227 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.767351 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.813860 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/3.log" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.815045 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/2.log" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.819128 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9" exitCode=1 Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.819166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.819209 4735 scope.go:117] "RemoveContainer" containerID="382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.821306 4735 scope.go:117] "RemoveContainer" containerID="c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9" Nov 22 08:04:18 crc kubenswrapper[4735]: E1122 08:04:18.821708 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.846804 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.863513 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.869931 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.869975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.869987 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.870009 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.870025 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.877729 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"2025-11-22T08:03:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d\\\\n2025-11-22T08:03:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d to /host/opt/cni/bin/\\\\n2025-11-22T08:03:27Z [verbose] multus-daemon started\\\\n2025-11-22T08:03:27Z [verbose] Readiness Indicator file check\\\\n2025-11-22T08:04:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:04:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.906924 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.923499 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.938565 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.961936 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"nternal-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress/router-internal-default_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1122 08:04:18.102104 6735 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:04:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.975807 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.975956 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.975980 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.976048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.976066 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:18Z","lastTransitionTime":"2025-11-22T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.976430 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:18 crc kubenswrapper[4735]: I1122 08:04:18.990078 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:18Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.002919 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.016696 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.026274 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.037028 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.052272 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.066975 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.078649 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.078684 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.078693 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.078706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.078760 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.079989 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.094251 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.106578 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.182189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.182237 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.182252 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.182270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.182282 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.263075 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:19 crc kubenswrapper[4735]: E1122 08:04:19.263213 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.291953 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.296909 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.296938 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.296945 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.296960 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.296968 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.327800 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.339262 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.351699 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.364089 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.377193 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.387825 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.403682 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.404528 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.404566 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.404586 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.404604 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.404619 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.416004 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.425593 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.436659 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.447609 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"2025-11-22T08:03:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d\\\\n2025-11-22T08:03:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d to /host/opt/cni/bin/\\\\n2025-11-22T08:03:27Z [verbose] multus-daemon started\\\\n2025-11-22T08:03:27Z [verbose] Readiness Indicator file check\\\\n2025-11-22T08:04:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:04:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.460671 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.472479 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.486288 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.505248 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://382485c1979a1f2654ac1d06ff6c49c32ca20b174f252b758085519e50ed9a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:03:52Z\\\",\\\"message\\\":\\\"1122 08:03:52.376727 6387 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 08:03:52.376871 6387 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 08:03:52.376883 6387 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 08:03:52.376888 6387 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1122 08:03:52.376895 6387 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 08:03:52.376901 6387 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 08:03:52.377054 6387 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1122 08:03:52.377067 6387 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377060 6387 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 08:03:52.377348 6387 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 08:03:52.377687 6387 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"nternal-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress/router-internal-default_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1122 08:04:18.102104 6735 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:04:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.506706 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.506735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.506744 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.506758 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.506767 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.517066 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.538044 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.609478 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.609518 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.609527 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.609541 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.609550 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.712793 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.712919 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.712946 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.713023 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.713048 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.814760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.815070 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.815082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.815099 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.815112 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.823856 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/3.log" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.827046 4735 scope.go:117] "RemoveContainer" containerID="c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9" Nov 22 08:04:19 crc kubenswrapper[4735]: E1122 08:04:19.827272 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.842277 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.857881 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.872406 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.887489 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.901082 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.918255 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.918326 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.918509 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.918602 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.918623 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:19Z","lastTransitionTime":"2025-11-22T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.919482 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.939452 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"2025-11-22T08:03:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d\\\\n2025-11-22T08:03:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d to /host/opt/cni/bin/\\\\n2025-11-22T08:03:27Z [verbose] multus-daemon started\\\\n2025-11-22T08:03:27Z [verbose] Readiness Indicator file check\\\\n2025-11-22T08:04:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:04:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.962146 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.980754 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:19 crc kubenswrapper[4735]: I1122 08:04:19.997276 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:19Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.017282 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"nternal-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress/router-internal-default_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1122 08:04:18.102104 6735 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:04:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:20Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.021438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.021493 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.021506 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.021523 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.021535 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.029507 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:20Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.049912 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:20Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.063669 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:20Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.081673 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:20Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.095105 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:20Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.108448 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:20Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.118791 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:20Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.124120 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.124168 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.124182 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.124202 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.124215 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.227703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.227782 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.227799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.227824 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.227843 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.262539 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.262710 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:20 crc kubenswrapper[4735]: E1122 08:04:20.262870 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.262902 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:20 crc kubenswrapper[4735]: E1122 08:04:20.263008 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:20 crc kubenswrapper[4735]: E1122 08:04:20.263109 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.330826 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.330898 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.330913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.330940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.330957 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.433525 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.433604 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.433628 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.433658 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.433676 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.536446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.536534 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.536551 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.536575 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.536591 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.639055 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.639094 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.639105 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.639120 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.639132 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.742028 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.742072 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.742083 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.742098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.742110 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.843955 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.844007 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.844019 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.844041 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.844053 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.947914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.947975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.947992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.948015 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:20 crc kubenswrapper[4735]: I1122 08:04:20.948033 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:20Z","lastTransitionTime":"2025-11-22T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.051563 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.051611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.051622 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.051638 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.051651 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.154397 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.154490 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.154508 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.154532 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.154549 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.258584 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.258662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.258686 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.258718 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.258790 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.263029 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:21 crc kubenswrapper[4735]: E1122 08:04:21.263291 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.362108 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.362147 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.362156 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.362172 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.362184 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.465359 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.465405 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.465417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.465435 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.465447 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.568417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.568485 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.568498 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.568514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.568525 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.671883 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.671938 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.671949 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.671965 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.671977 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.774200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.774253 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.774266 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.774293 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.774305 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.876635 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.876669 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.876678 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.876692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.876702 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.979846 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.979912 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.979935 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.979963 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:21 crc kubenswrapper[4735]: I1122 08:04:21.979984 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:21Z","lastTransitionTime":"2025-11-22T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.082758 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.082798 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.082809 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.082826 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.082838 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.186214 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.186280 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.186297 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.186323 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.186341 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.263108 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.263162 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:22 crc kubenswrapper[4735]: E1122 08:04:22.263305 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.263393 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:22 crc kubenswrapper[4735]: E1122 08:04:22.263588 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:22 crc kubenswrapper[4735]: E1122 08:04:22.263921 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.278768 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.290749 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.290846 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.290873 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.290913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.290951 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.394206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.394268 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.394286 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.394309 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.394329 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.497690 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.497770 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.497804 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.497832 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.497853 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.601222 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.601295 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.601313 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.601336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.601353 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.704785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.704852 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.704863 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.704881 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.704894 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.808140 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.808183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.808196 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.808215 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.808229 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.910371 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.910406 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.910417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.910432 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:22 crc kubenswrapper[4735]: I1122 08:04:22.910443 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:22Z","lastTransitionTime":"2025-11-22T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.012496 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.012547 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.012562 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.012587 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.012604 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.115836 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.115890 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.115902 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.115917 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.115925 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.218692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.218781 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.218792 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.218810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.218822 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.263757 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:23 crc kubenswrapper[4735]: E1122 08:04:23.263880 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.321545 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.321598 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.321615 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.321637 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.321654 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.424392 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.424446 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.424489 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.424511 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.424526 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.527088 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.527151 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.527168 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.527193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.527219 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.630557 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.630617 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.630633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.630653 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.630666 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.733308 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.733425 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.733439 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.733481 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.733497 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.838015 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.838077 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.838092 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.838112 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.838127 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.940640 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.940701 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.940718 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.940737 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:23 crc kubenswrapper[4735]: I1122 08:04:23.940751 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:23Z","lastTransitionTime":"2025-11-22T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.005564 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.005643 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.005777 4735 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.005809 4735 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.005833 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.005816599 +0000 UTC m=+149.610155224 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.005934 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.005910722 +0000 UTC m=+149.610249337 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.044174 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.044218 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.044231 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.044249 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.044262 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.107289 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.107511 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.107577 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.107532083 +0000 UTC m=+149.711870718 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.107695 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.107707 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.107897 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.107782 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.107924 4735 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.107938 4735 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.107952 4735 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.107989 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.107970875 +0000 UTC m=+149.712309500 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.108015 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.108004306 +0000 UTC m=+149.712343121 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.148075 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.148133 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.148158 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.148183 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.148200 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.251257 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.251348 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.251362 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.251412 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.251433 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.262771 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.262803 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.262958 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.263043 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.263218 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:24 crc kubenswrapper[4735]: E1122 08:04:24.263396 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.353866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.353905 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.353921 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.353936 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.353946 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.457897 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.457990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.458018 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.458054 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.458082 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.562136 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.562216 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.562247 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.562277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.562298 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.665514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.665594 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.665604 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.665630 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.665649 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.768743 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.768790 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.768800 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.768814 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.768824 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.871475 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.871537 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.871547 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.871569 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.871581 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.974261 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.974327 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.974345 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.974374 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:24 crc kubenswrapper[4735]: I1122 08:04:24.974391 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:24Z","lastTransitionTime":"2025-11-22T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.077927 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.078009 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.078035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.078066 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.078084 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.180977 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.181043 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.181090 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.181111 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.181126 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.262396 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:25 crc kubenswrapper[4735]: E1122 08:04:25.262561 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.283455 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.283522 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.283537 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.283554 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.283567 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.386932 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.386997 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.387012 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.387034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.387053 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.489724 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.489775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.489787 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.489805 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.489818 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.592784 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.592842 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.592854 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.592913 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.592926 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.696399 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.696482 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.696496 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.696518 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.696533 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.800149 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.800205 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.800217 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.800240 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.800251 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.903169 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.903230 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.903245 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.903265 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:25 crc kubenswrapper[4735]: I1122 08:04:25.903278 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:25Z","lastTransitionTime":"2025-11-22T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.006628 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.006708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.006729 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.006764 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.006788 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.110444 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.110563 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.110584 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.110615 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.110640 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.214185 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.214245 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.214262 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.214286 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.214306 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.263139 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.263193 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.263139 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:26 crc kubenswrapper[4735]: E1122 08:04:26.263310 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:26 crc kubenswrapper[4735]: E1122 08:04:26.263439 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:26 crc kubenswrapper[4735]: E1122 08:04:26.263593 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.317507 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.317556 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.317574 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.317596 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.317615 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.421140 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.421220 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.421248 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.421278 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.421300 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.524068 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.524098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.524108 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.524119 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.524127 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.626672 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.626719 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.626731 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.626749 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.626762 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.729972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.730020 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.730033 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.730053 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.730064 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.832799 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.832843 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.832857 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.832878 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.832891 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.936083 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.936153 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.936177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.936208 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:26 crc kubenswrapper[4735]: I1122 08:04:26.936269 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:26Z","lastTransitionTime":"2025-11-22T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.039095 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.039163 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.039181 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.039206 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.039224 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.142217 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.142291 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.142313 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.142339 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.142357 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.246135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.246188 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.246202 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.246224 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.246241 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.262552 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:27 crc kubenswrapper[4735]: E1122 08:04:27.262808 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.348885 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.348967 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.348983 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.349012 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.349030 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.452151 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.452211 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.452231 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.452260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.452282 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.555599 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.555694 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.555728 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.555760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.555785 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.658974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.659031 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.659046 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.659071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.659090 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.762260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.762338 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.762356 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.762386 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.762405 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.865640 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.865703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.865721 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.865746 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.865764 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.969059 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.969122 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.969135 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.969152 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:27 crc kubenswrapper[4735]: I1122 08:04:27.969164 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:27Z","lastTransitionTime":"2025-11-22T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.058345 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.059160 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.059202 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.059235 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.059257 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.077163 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.084261 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.084306 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.084321 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.084338 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.084352 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.102181 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.108872 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.108943 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.108965 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.108993 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.109011 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.132692 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.137862 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.137930 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.137950 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.137974 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.137993 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.161315 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.166641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.166727 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.166754 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.166785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.166805 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.185936 4735 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f042384-cd21-4499-89ef-6191a35371a5\\\",\\\"systemUUID\\\":\\\"461f179d-030f-4c48-b90b-03251dc5d5b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:28Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.186163 4735 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.188514 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.188594 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.188608 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.188655 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.188676 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.262859 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.262857 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.263034 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.262877 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.263199 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:28 crc kubenswrapper[4735]: E1122 08:04:28.263412 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.291725 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.291973 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.291986 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.292007 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.292020 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.394956 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.394998 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.395033 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.395497 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.395518 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.499152 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.499218 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.499242 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.499270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.499291 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.601843 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.601926 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.601949 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.601977 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.601999 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.704937 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.704996 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.705013 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.705033 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.705048 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.808075 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.808122 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.808134 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.808150 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.808163 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.910601 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.910644 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.910655 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.910671 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:28 crc kubenswrapper[4735]: I1122 08:04:28.910682 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:28Z","lastTransitionTime":"2025-11-22T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.013127 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.013181 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.013192 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.013210 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.013224 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.116422 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.116674 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.116703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.116728 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.116750 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.219685 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.220319 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.220608 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.220801 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.220901 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.263825 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:29 crc kubenswrapper[4735]: E1122 08:04:29.264784 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.294090 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f39e5f7-6ef4-4c58-87d2-5732362e96ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f6b43dc842b3ebdfd382dd8ac547042cc47dcb404ce438f250982349ebd01c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d04030ec9a23d9fae2d28b709bed8b16e9f8bb429e3bdac62184a3b7e6720\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd1fb496cc9fc5aad906e6ddfc327de1e0ac3176b04f65af13edde71deca15a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c8fab9f0f447d594281d779abf068382cac2348dad7964ae80d3ccdd145bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a507a23b493eb6f8bd57b1d678461bc81db7b93a02698ae95147cbb3b06f81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f76fb607600b09c20f1f4e4e3d154ec75904f1249ca3a87cc925c236654d3a64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5cff0c85f46c66cc1c1920dcfd372b1147f4a5d0ca23e13feeb4a87a8d0ec92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2933ee447ea715084de60ec8623b07d445ab63dffca92288c343ff0ea6eb721\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.314446 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78ca3124-02cf-4142-a42e-4b3e70eb2882\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff0eef744c658b86a87105d6143a594fe4f25f08cca66302fec7828bd9aa262\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7762816b69abef500f9fe0791231bba8d156b088996c11d6926814ff16e86b2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b121a579b26c0d732e4aefdc92689fb6c82fa2b019ded043b776993324af91c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dad0d2be20d22a8e9610ec18bfd722ac09269b5590ad142caee16d2cef4fa088\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.324665 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.324730 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.324750 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.324775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.324797 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.334153 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7945a41-afa2-43b7-95a2-085e3b62e64b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b35ec4d0cf2e5a33f2fd4631ff5a3ba7db390c7a79307b966d71c2630356ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t88z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dt2lg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.355145 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2719040d-3088-41c7-8f16-5508d78669e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:18Z\\\",\\\"message\\\":\\\"nternal-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress/router-internal-default_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1122 08:04:18.102104 6735 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:04:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvx5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pl875\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.372011 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ac395c9-165f-473f-a4da-7655ae4e89a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://027ef2d91b9eaf985c71ba22e445a1ec7c5ed737fd32ecedf7d9413859f73a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f355e529d5db7c9e1aa71f2ba1fab45002f1d553ccfe3f183badf135b52112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cwlr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.386585 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51cc79a45575114c1e839066e9f3dbe2263346385d0f9f466ab1680edceb93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.400920 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-x2srr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78c491a2-c7e5-4084-a48b-c6dd8163b467\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2645c7b19a0d24ee0c6068d5f20c87b62aafa4dc4c712dbeea248d372098c230\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64hff\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x2srr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.427209 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-25pl8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5003f416-41bd-48b2-8965-292d0f933500\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1626037233d9218f6cee9e70e9dcf23b98949ccacead18ca1903b21448c6f79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de3040e542b379ac21c250a5b3312fe7fc81ee2de1b403b9a2d6e117df5c7887\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfd0eafe9e0aac38a12d1f2370421223dcf20742749c85c247b8cab5949e3b76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c3eadca16481b36f5147bf16bb1210ada2cda1d606fa3d19d867c4a084b84e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa4ee3b7e2284255b04df05be88ab996670b276f1eac0d50166b6a2a76b69d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://581c57b4034b2e6db0c7d339417e87bd26fc0efa825b1c52402d3d7fa2d58816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3485f315d681833ab89e9cc8709cf8c6ce2841ba262418f653c1ac132bd1ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgldm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-25pl8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.428113 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.428165 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.428174 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.428189 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.428198 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.446210 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vvf57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqpfj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vvf57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.464389 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57a1d7c8ab7178004f3aac53272dab7cfe56b6fd6911b575f23e954fd460a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5108b6fe5835d2bc2cde227d1f464a66aa5ef66cf1bc1d06eb7c65ca28a7b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.478508 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l2rhr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646e7dd4-85e4-448c-bdea-4e5870f5530f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc7b350edf6b11ec58d5c89f49f45c2748ad99de4c2bdb33b0c152adea57b551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4ns7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l2rhr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.497754 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4eb581a4-ac69-423f-be54-a478c351e493\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03bc10ac96df86b5aa93e10f2dfd9153d08eb85a2d8b17efdf729f0360bb5e45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6a4d2467c841f0bd41aa23cf9c18b55786412f686d69606c6e947205e3e4d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6a4d2467c841f0bd41aa23cf9c18b55786412f686d69606c6e947205e3e4d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.520240 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f6104e-8e91-46ba-8f4c-2dceb81ae347\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd6dda2800a7b4de6bbaa2a59de60b7fa5986e90b4399e1e7e45eb183f3c301\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://667c19d95a0a07c22574e539aa595832643f9df5ecc0a838aa4a838be9b8062f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5caa37c847d9460d751e3763728d3ec57423cfc9a0d263d875ceda3201791ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b06b18f98eda17ed8af9cddb401392f6d1c00ed81fc7af433df39e2a7a4ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.531578 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.531621 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.531633 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.531651 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.531663 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.541666 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d33b5c53d1d7a29148974f3288c98171bfecc88e664df2874e81b9089d635d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.562438 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.577596 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.596355 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b9debc3-c362-4bd1-8f64-f97d8761efe7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af230ecef757389a24f80b0ea2e1a34f5f8994e85f42cc5bff55a4a73a49603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bc44e8c4eb2bbc47a257627249957efede6d498128af7d3a6f8caf1d101f75e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d6a3d5bb14b9260012ccd81ee68a17b2211e384f844a30a578989118f18ecc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac3d6ed2a90ea4b85f0d06925509a35d9e0e717c8c957d2990cd7d5e28cb520d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33a3d41790f073a3265259ff7db05338dfce952e1bff7af0dc4af3d48582dc37\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 08:03:20.052264 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 08:03:20.052359 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 08:03:20.053178 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1268944283/tls.crt::/tmp/serving-cert-1268944283/tls.key\\\\\\\"\\\\nI1122 08:03:20.331975 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 08:03:20.335638 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 08:03:20.335668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 08:03:20.335696 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 08:03:20.335703 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 08:03:20.341181 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 08:03:20.341215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341222 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 08:03:20.341229 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 08:03:20.341233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 08:03:20.341237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 08:03:20.341241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1122 08:03:20.341247 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1122 08:03:20.344755 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://753474ada70a40a033aa1a493da60724b149530e37ad928349cda9b75dfc6963\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:03:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b96a12445b8499d7162ac29a40eae4d7941abd1b8cfaf2e3fa57de970339b07b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:02:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.611411 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:20Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.626097 4735 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-85z2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61dbab7f-c903-4b55-9d90-aacfd12f7f0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T08:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T08:04:12Z\\\",\\\"message\\\":\\\"2025-11-22T08:03:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d\\\\n2025-11-22T08:03:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71cc430a-34eb-4767-837d-ad4a3e06f16d to /host/opt/cni/bin/\\\\n2025-11-22T08:03:27Z [verbose] multus-daemon started\\\\n2025-11-22T08:03:27Z [verbose] Readiness Indicator file check\\\\n2025-11-22T08:04:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T08:03:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T08:04:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sz5mx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T08:03:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-85z2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T08:04:29Z is after 2025-08-24T17:21:41Z" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.634024 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.634090 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.634106 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.634127 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.634142 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.737430 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.737552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.737579 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.737610 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.737632 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.840831 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.840896 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.840906 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.840926 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.840940 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.944624 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.944969 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.944979 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.944999 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:29 crc kubenswrapper[4735]: I1122 08:04:29.945010 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:29Z","lastTransitionTime":"2025-11-22T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.047708 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.047751 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.047760 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.047774 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.047784 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.150501 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.150571 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.150581 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.150603 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.150616 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.254491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.254564 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.254583 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.254608 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.254626 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.262793 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.262829 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:30 crc kubenswrapper[4735]: E1122 08:04:30.262965 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.263040 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:30 crc kubenswrapper[4735]: E1122 08:04:30.263331 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:30 crc kubenswrapper[4735]: E1122 08:04:30.263791 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.358312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.358404 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.358496 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.358584 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.358601 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.462251 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.462343 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.462393 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.462416 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.462435 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.565675 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.565735 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.565753 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.565778 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.565795 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.670006 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.670089 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.670108 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.670692 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.670758 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.773818 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.773868 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.773912 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.773929 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.773940 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.876516 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.876569 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.876586 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.876608 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.876626 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.981990 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.982057 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.982071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.982098 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:30 crc kubenswrapper[4735]: I1122 08:04:30.982113 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:30Z","lastTransitionTime":"2025-11-22T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.085944 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.086014 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.086036 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.086065 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.086087 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.189342 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.189417 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.189436 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.189495 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.189519 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.263238 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:31 crc kubenswrapper[4735]: E1122 08:04:31.263387 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.264925 4735 scope.go:117] "RemoveContainer" containerID="c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9" Nov 22 08:04:31 crc kubenswrapper[4735]: E1122 08:04:31.265163 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.291895 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.291929 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.291942 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.291975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.291988 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.395685 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.395766 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.395785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.395812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.395833 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.499295 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.499360 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.499378 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.499403 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.499421 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.629856 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.629922 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.629935 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.629958 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.629974 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.733133 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.733193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.733213 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.733239 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.733257 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.836210 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.836260 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.836273 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.836290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.836301 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.939115 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.939177 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.939193 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.939221 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:31 crc kubenswrapper[4735]: I1122 08:04:31.939239 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:31Z","lastTransitionTime":"2025-11-22T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.042566 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.043034 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.043296 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.043958 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.044207 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.147361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.147742 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.147925 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.148138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.148302 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.250615 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.250688 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.250705 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.250733 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.250752 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.263243 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.263308 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.263397 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:32 crc kubenswrapper[4735]: E1122 08:04:32.263736 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:32 crc kubenswrapper[4735]: E1122 08:04:32.263856 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:32 crc kubenswrapper[4735]: E1122 08:04:32.264085 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.354427 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.354510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.354523 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.354557 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.354575 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.458701 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.458775 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.458796 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.458827 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.458849 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.562061 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.562132 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.562146 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.562170 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.562187 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.665172 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.665277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.665300 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.665326 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.665343 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.768947 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.768992 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.769004 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.769020 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.769036 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.871981 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.872041 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.872055 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.872082 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.872095 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.975785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.975864 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.975889 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.975919 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:32 crc kubenswrapper[4735]: I1122 08:04:32.975941 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:32Z","lastTransitionTime":"2025-11-22T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.079179 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.079229 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.079240 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.079261 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.079273 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.182732 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.182812 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.182835 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.182866 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.182888 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.262896 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:33 crc kubenswrapper[4735]: E1122 08:04:33.263060 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.286369 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.286438 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.286488 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.286512 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.286530 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.390380 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.390894 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.391231 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.391614 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.391765 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.495312 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.495371 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.495389 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.495414 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.495434 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.599777 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.600318 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.600717 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.601134 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.601523 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.705451 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.705521 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.705531 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.705552 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.705565 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.809768 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.809821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.809839 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.809867 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.809886 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.912714 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.913197 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.913286 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.913377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:33 crc kubenswrapper[4735]: I1122 08:04:33.913536 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:33Z","lastTransitionTime":"2025-11-22T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.018050 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.018121 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.018138 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.018165 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.018184 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.122215 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.122303 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.122323 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.122356 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.122380 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.225377 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.225427 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.225445 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.225513 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.225535 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.262810 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.262812 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:34 crc kubenswrapper[4735]: E1122 08:04:34.262995 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:34 crc kubenswrapper[4735]: E1122 08:04:34.263297 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.262847 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:34 crc kubenswrapper[4735]: E1122 08:04:34.263519 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.328443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.328542 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.328560 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.328586 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.328604 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.432156 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.432220 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.432238 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.432277 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.432304 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.535810 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.535882 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.535897 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.535917 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.535929 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.639947 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.640013 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.640026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.640048 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.640061 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.743729 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.743780 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.743792 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.743811 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.743823 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.847318 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.847400 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.847419 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.847450 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.847501 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.951836 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.951914 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.951938 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.951972 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:34 crc kubenswrapper[4735]: I1122 08:04:34.951994 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:34Z","lastTransitionTime":"2025-11-22T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.054823 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.054904 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.054922 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.054946 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.054962 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.157292 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.157351 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.157369 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.157392 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.157410 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.260767 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.260834 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.260857 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.260885 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.260906 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.262694 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:35 crc kubenswrapper[4735]: E1122 08:04:35.263004 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.364662 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.364733 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.364756 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.364790 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.364813 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.467524 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.467837 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.467971 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.468109 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.468289 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.576724 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.576761 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.576771 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.576785 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.576795 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.679962 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.680026 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.680045 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.680069 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.680129 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.783954 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.784022 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.784071 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.784097 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.784114 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.886678 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.886738 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.886756 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.886781 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.886799 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.989412 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.989501 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.989515 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.989538 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:35 crc kubenswrapper[4735]: I1122 08:04:35.989553 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:35Z","lastTransitionTime":"2025-11-22T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.092141 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.092232 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.092253 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.092286 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.092308 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.195281 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.195676 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.195821 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.196042 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.196202 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.262680 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.262746 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:36 crc kubenswrapper[4735]: E1122 08:04:36.262844 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.262704 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:36 crc kubenswrapper[4735]: E1122 08:04:36.263323 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:36 crc kubenswrapper[4735]: E1122 08:04:36.263576 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.299270 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.299313 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.299321 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.299365 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.299376 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.402726 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.403200 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.403388 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.403603 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.403821 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.507363 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.507432 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.507452 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.507510 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.507532 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.610596 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.610657 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.610676 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.610703 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.610724 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.714241 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.714296 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.714313 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.714336 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.714352 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.817790 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.817940 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.818035 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.818063 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.818080 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.921344 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.921443 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.921491 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.921523 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:36 crc kubenswrapper[4735]: I1122 08:04:36.921546 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:36Z","lastTransitionTime":"2025-11-22T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.024424 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.024522 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.024546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.024574 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.024594 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.127993 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.128094 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.128114 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.128144 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.128162 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.231668 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.231748 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.231772 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.231804 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.231832 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.263169 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:37 crc kubenswrapper[4735]: E1122 08:04:37.263398 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.334733 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.334817 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.334833 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.334857 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.334875 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.438573 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.438641 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.438667 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.438697 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.438719 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.542290 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.542386 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.542415 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.542449 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.542512 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.645743 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.645838 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.645861 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.645893 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.645915 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.749532 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.749602 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.749618 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.749643 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.749659 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.852975 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.853036 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.853052 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.853079 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.853098 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.956019 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.956074 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.956083 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.956101 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:37 crc kubenswrapper[4735]: I1122 08:04:37.956113 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:37Z","lastTransitionTime":"2025-11-22T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.059053 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.059322 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.059361 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.059393 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.059415 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:38Z","lastTransitionTime":"2025-11-22T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.163166 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.163241 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.163265 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.163292 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.163315 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:38Z","lastTransitionTime":"2025-11-22T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.262241 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.262302 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.262307 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:38 crc kubenswrapper[4735]: E1122 08:04:38.262413 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:38 crc kubenswrapper[4735]: E1122 08:04:38.262560 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:38 crc kubenswrapper[4735]: E1122 08:04:38.262711 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.267546 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.267598 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.267611 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.267628 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.267644 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:38Z","lastTransitionTime":"2025-11-22T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.371318 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.371569 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.371634 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.372051 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.372264 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:38Z","lastTransitionTime":"2025-11-22T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.476296 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.476363 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.476381 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.476403 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.476421 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:38Z","lastTransitionTime":"2025-11-22T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.514855 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.514912 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.514928 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.514954 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.514970 4735 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T08:04:38Z","lastTransitionTime":"2025-11-22T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.578431 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2"] Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.578947 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.583032 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.583032 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.583925 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.583930 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.606018 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/039a3e4f-2c26-460c-a724-2f5bb533f9e2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.606089 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039a3e4f-2c26-460c-a724-2f5bb533f9e2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.606114 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/039a3e4f-2c26-460c-a724-2f5bb533f9e2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.606141 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/039a3e4f-2c26-460c-a724-2f5bb533f9e2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.606199 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/039a3e4f-2c26-460c-a724-2f5bb533f9e2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.611768 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cwlr6" podStartSLOduration=77.611747443 podStartE2EDuration="1m17.611747443s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.597843334 +0000 UTC m=+100.202182059" watchObservedRunningTime="2025-11-22 08:04:38.611747443 +0000 UTC m=+100.216086058" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.643544 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=78.643529259 podStartE2EDuration="1m18.643529259s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.643495288 +0000 UTC m=+100.247833933" watchObservedRunningTime="2025-11-22 08:04:38.643529259 +0000 UTC m=+100.247867864" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.670286 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=50.670271228 podStartE2EDuration="50.670271228s" podCreationTimestamp="2025-11-22 08:03:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.659405363 +0000 UTC m=+100.263743998" watchObservedRunningTime="2025-11-22 08:04:38.670271228 +0000 UTC m=+100.274609833" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.670538 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podStartSLOduration=78.670533136 podStartE2EDuration="1m18.670533136s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.670097403 +0000 UTC m=+100.274436018" watchObservedRunningTime="2025-11-22 08:04:38.670533136 +0000 UTC m=+100.274871741" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.707488 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/039a3e4f-2c26-460c-a724-2f5bb533f9e2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.707548 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/039a3e4f-2c26-460c-a724-2f5bb533f9e2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.707574 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/039a3e4f-2c26-460c-a724-2f5bb533f9e2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.707608 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/039a3e4f-2c26-460c-a724-2f5bb533f9e2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.707628 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039a3e4f-2c26-460c-a724-2f5bb533f9e2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.707634 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/039a3e4f-2c26-460c-a724-2f5bb533f9e2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.707774 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/039a3e4f-2c26-460c-a724-2f5bb533f9e2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.708336 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/039a3e4f-2c26-460c-a724-2f5bb533f9e2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.713759 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039a3e4f-2c26-460c-a724-2f5bb533f9e2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.724052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/039a3e4f-2c26-460c-a724-2f5bb533f9e2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9s7w2\" (UID: \"039a3e4f-2c26-460c-a724-2f5bb533f9e2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.725368 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-x2srr" podStartSLOduration=78.7253479 podStartE2EDuration="1m18.7253479s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.725233887 +0000 UTC m=+100.329572482" watchObservedRunningTime="2025-11-22 08:04:38.7253479 +0000 UTC m=+100.329686505" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.751419 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-25pl8" podStartSLOduration=78.75138459 podStartE2EDuration="1m18.75138459s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.739658751 +0000 UTC m=+100.343997356" watchObservedRunningTime="2025-11-22 08:04:38.75138459 +0000 UTC m=+100.355723205" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.786490 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-l2rhr" podStartSLOduration=78.786448527 podStartE2EDuration="1m18.786448527s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.785897222 +0000 UTC m=+100.390235847" watchObservedRunningTime="2025-11-22 08:04:38.786448527 +0000 UTC m=+100.390787132" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.800161 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=16.80013736 podStartE2EDuration="16.80013736s" podCreationTimestamp="2025-11-22 08:04:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.80013704 +0000 UTC m=+100.404475645" watchObservedRunningTime="2025-11-22 08:04:38.80013736 +0000 UTC m=+100.404475965" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.814487 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.814449901 podStartE2EDuration="1m18.814449901s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.814331697 +0000 UTC m=+100.418670302" watchObservedRunningTime="2025-11-22 08:04:38.814449901 +0000 UTC m=+100.418788506" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.858254 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.858235575 podStartE2EDuration="1m17.858235575s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.857416942 +0000 UTC m=+100.461755557" watchObservedRunningTime="2025-11-22 08:04:38.858235575 +0000 UTC m=+100.462574180" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.889195 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-85z2h" podStartSLOduration=78.889172749 podStartE2EDuration="1m18.889172749s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:38.887902644 +0000 UTC m=+100.492241279" watchObservedRunningTime="2025-11-22 08:04:38.889172749 +0000 UTC m=+100.493511374" Nov 22 08:04:38 crc kubenswrapper[4735]: I1122 08:04:38.903449 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" Nov 22 08:04:39 crc kubenswrapper[4735]: I1122 08:04:39.213593 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:39 crc kubenswrapper[4735]: E1122 08:04:39.214081 4735 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:04:39 crc kubenswrapper[4735]: E1122 08:04:39.214134 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs podName:e9170f64-5b0d-4b2b-99ce-7ecb9f567620 nodeName:}" failed. No retries permitted until 2025-11-22 08:05:43.214117721 +0000 UTC m=+164.818456336 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs") pod "network-metrics-daemon-vvf57" (UID: "e9170f64-5b0d-4b2b-99ce-7ecb9f567620") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 08:04:39 crc kubenswrapper[4735]: I1122 08:04:39.262767 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:39 crc kubenswrapper[4735]: E1122 08:04:39.265110 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:39 crc kubenswrapper[4735]: I1122 08:04:39.892615 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" event={"ID":"039a3e4f-2c26-460c-a724-2f5bb533f9e2","Type":"ContainerStarted","Data":"ff19b0a8144cb52089a4f8c9064c410885cefc8f0c8b6196f2375cbf381fabf2"} Nov 22 08:04:39 crc kubenswrapper[4735]: I1122 08:04:39.892672 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" event={"ID":"039a3e4f-2c26-460c-a724-2f5bb533f9e2","Type":"ContainerStarted","Data":"29ddb5d64f9a2fc2a5bd47c2ee5e203c68175915db6f10e807db78cb3ca3d3d8"} Nov 22 08:04:39 crc kubenswrapper[4735]: I1122 08:04:39.911426 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9s7w2" podStartSLOduration=79.911407027 podStartE2EDuration="1m19.911407027s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:04:39.910919503 +0000 UTC m=+101.515258108" watchObservedRunningTime="2025-11-22 08:04:39.911407027 +0000 UTC m=+101.515745632" Nov 22 08:04:40 crc kubenswrapper[4735]: I1122 08:04:40.262958 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:40 crc kubenswrapper[4735]: I1122 08:04:40.263186 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:40 crc kubenswrapper[4735]: I1122 08:04:40.263405 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:40 crc kubenswrapper[4735]: E1122 08:04:40.263404 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:40 crc kubenswrapper[4735]: E1122 08:04:40.263572 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:40 crc kubenswrapper[4735]: E1122 08:04:40.263778 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:41 crc kubenswrapper[4735]: I1122 08:04:41.263106 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:41 crc kubenswrapper[4735]: E1122 08:04:41.263513 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:42 crc kubenswrapper[4735]: I1122 08:04:42.263124 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:42 crc kubenswrapper[4735]: I1122 08:04:42.263164 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:42 crc kubenswrapper[4735]: I1122 08:04:42.263164 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:42 crc kubenswrapper[4735]: E1122 08:04:42.263490 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:42 crc kubenswrapper[4735]: E1122 08:04:42.263576 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:42 crc kubenswrapper[4735]: E1122 08:04:42.263872 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:43 crc kubenswrapper[4735]: I1122 08:04:43.262882 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:43 crc kubenswrapper[4735]: E1122 08:04:43.263087 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:44 crc kubenswrapper[4735]: I1122 08:04:44.262326 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:44 crc kubenswrapper[4735]: I1122 08:04:44.262356 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:44 crc kubenswrapper[4735]: I1122 08:04:44.262501 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:44 crc kubenswrapper[4735]: E1122 08:04:44.262496 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:44 crc kubenswrapper[4735]: E1122 08:04:44.262643 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:44 crc kubenswrapper[4735]: E1122 08:04:44.262700 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:45 crc kubenswrapper[4735]: I1122 08:04:45.262849 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:45 crc kubenswrapper[4735]: E1122 08:04:45.263053 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:46 crc kubenswrapper[4735]: I1122 08:04:46.262763 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:46 crc kubenswrapper[4735]: I1122 08:04:46.262785 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:46 crc kubenswrapper[4735]: E1122 08:04:46.262887 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:46 crc kubenswrapper[4735]: I1122 08:04:46.262763 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:46 crc kubenswrapper[4735]: E1122 08:04:46.263046 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:46 crc kubenswrapper[4735]: E1122 08:04:46.263243 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:46 crc kubenswrapper[4735]: I1122 08:04:46.264327 4735 scope.go:117] "RemoveContainer" containerID="c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9" Nov 22 08:04:46 crc kubenswrapper[4735]: E1122 08:04:46.264608 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:04:47 crc kubenswrapper[4735]: I1122 08:04:47.263039 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:47 crc kubenswrapper[4735]: E1122 08:04:47.263213 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:48 crc kubenswrapper[4735]: I1122 08:04:48.263040 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:48 crc kubenswrapper[4735]: I1122 08:04:48.263044 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:48 crc kubenswrapper[4735]: E1122 08:04:48.263292 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:48 crc kubenswrapper[4735]: E1122 08:04:48.263358 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:48 crc kubenswrapper[4735]: I1122 08:04:48.263087 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:48 crc kubenswrapper[4735]: E1122 08:04:48.263445 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:49 crc kubenswrapper[4735]: I1122 08:04:49.263075 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:49 crc kubenswrapper[4735]: E1122 08:04:49.264869 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:50 crc kubenswrapper[4735]: I1122 08:04:50.262746 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:50 crc kubenswrapper[4735]: I1122 08:04:50.262798 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:50 crc kubenswrapper[4735]: E1122 08:04:50.263112 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:50 crc kubenswrapper[4735]: I1122 08:04:50.263419 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:50 crc kubenswrapper[4735]: E1122 08:04:50.263559 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:50 crc kubenswrapper[4735]: E1122 08:04:50.263812 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:51 crc kubenswrapper[4735]: I1122 08:04:51.262732 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:51 crc kubenswrapper[4735]: E1122 08:04:51.263006 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:52 crc kubenswrapper[4735]: I1122 08:04:52.262545 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:52 crc kubenswrapper[4735]: I1122 08:04:52.262573 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:52 crc kubenswrapper[4735]: I1122 08:04:52.262682 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:52 crc kubenswrapper[4735]: E1122 08:04:52.262922 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:52 crc kubenswrapper[4735]: E1122 08:04:52.263012 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:52 crc kubenswrapper[4735]: E1122 08:04:52.263220 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:53 crc kubenswrapper[4735]: I1122 08:04:53.262971 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:53 crc kubenswrapper[4735]: E1122 08:04:53.263223 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:54 crc kubenswrapper[4735]: I1122 08:04:54.262254 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:54 crc kubenswrapper[4735]: E1122 08:04:54.262415 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:54 crc kubenswrapper[4735]: I1122 08:04:54.262522 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:54 crc kubenswrapper[4735]: E1122 08:04:54.262736 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:54 crc kubenswrapper[4735]: I1122 08:04:54.262911 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:54 crc kubenswrapper[4735]: E1122 08:04:54.263168 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:55 crc kubenswrapper[4735]: I1122 08:04:55.263008 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:55 crc kubenswrapper[4735]: E1122 08:04:55.263198 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:56 crc kubenswrapper[4735]: I1122 08:04:56.262954 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:56 crc kubenswrapper[4735]: I1122 08:04:56.262968 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:56 crc kubenswrapper[4735]: E1122 08:04:56.263180 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:56 crc kubenswrapper[4735]: I1122 08:04:56.262969 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:56 crc kubenswrapper[4735]: E1122 08:04:56.263336 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:56 crc kubenswrapper[4735]: E1122 08:04:56.263422 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:57 crc kubenswrapper[4735]: I1122 08:04:57.262548 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:57 crc kubenswrapper[4735]: E1122 08:04:57.262737 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:57 crc kubenswrapper[4735]: I1122 08:04:57.263995 4735 scope.go:117] "RemoveContainer" containerID="c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9" Nov 22 08:04:57 crc kubenswrapper[4735]: E1122 08:04:57.264456 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pl875_openshift-ovn-kubernetes(2719040d-3088-41c7-8f16-5508d78669e2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.262241 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.262283 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:04:58 crc kubenswrapper[4735]: E1122 08:04:58.262382 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.262449 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:04:58 crc kubenswrapper[4735]: E1122 08:04:58.262508 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:04:58 crc kubenswrapper[4735]: E1122 08:04:58.262745 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.969821 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/1.log" Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.970794 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/0.log" Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.970864 4735 generic.go:334] "Generic (PLEG): container finished" podID="61dbab7f-c903-4b55-9d90-aacfd12f7f0f" containerID="a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73" exitCode=1 Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.970906 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-85z2h" event={"ID":"61dbab7f-c903-4b55-9d90-aacfd12f7f0f","Type":"ContainerDied","Data":"a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73"} Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.970952 4735 scope.go:117] "RemoveContainer" containerID="04c02fce4d881f6c7ac68174db42c0ada619ed1f86552ba86582b9995f3bc5f5" Nov 22 08:04:58 crc kubenswrapper[4735]: I1122 08:04:58.971660 4735 scope.go:117] "RemoveContainer" containerID="a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73" Nov 22 08:04:58 crc kubenswrapper[4735]: E1122 08:04:58.972185 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-85z2h_openshift-multus(61dbab7f-c903-4b55-9d90-aacfd12f7f0f)\"" pod="openshift-multus/multus-85z2h" podUID="61dbab7f-c903-4b55-9d90-aacfd12f7f0f" Nov 22 08:04:59 crc kubenswrapper[4735]: I1122 08:04:59.262990 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:04:59 crc kubenswrapper[4735]: E1122 08:04:59.264106 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:04:59 crc kubenswrapper[4735]: E1122 08:04:59.291251 4735 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 22 08:04:59 crc kubenswrapper[4735]: E1122 08:04:59.351082 4735 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:04:59 crc kubenswrapper[4735]: I1122 08:04:59.976369 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/1.log" Nov 22 08:05:00 crc kubenswrapper[4735]: I1122 08:05:00.262419 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:00 crc kubenswrapper[4735]: I1122 08:05:00.262571 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:00 crc kubenswrapper[4735]: I1122 08:05:00.262419 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:00 crc kubenswrapper[4735]: E1122 08:05:00.262671 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:05:00 crc kubenswrapper[4735]: E1122 08:05:00.262881 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:05:00 crc kubenswrapper[4735]: E1122 08:05:00.263042 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:05:01 crc kubenswrapper[4735]: I1122 08:05:01.263600 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:01 crc kubenswrapper[4735]: E1122 08:05:01.263873 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:05:02 crc kubenswrapper[4735]: I1122 08:05:02.262560 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:02 crc kubenswrapper[4735]: I1122 08:05:02.262658 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:02 crc kubenswrapper[4735]: I1122 08:05:02.262562 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:02 crc kubenswrapper[4735]: E1122 08:05:02.262698 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:05:02 crc kubenswrapper[4735]: E1122 08:05:02.262847 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:05:02 crc kubenswrapper[4735]: E1122 08:05:02.262884 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:05:03 crc kubenswrapper[4735]: I1122 08:05:03.263392 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:03 crc kubenswrapper[4735]: E1122 08:05:03.263594 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:05:04 crc kubenswrapper[4735]: I1122 08:05:04.262871 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:04 crc kubenswrapper[4735]: I1122 08:05:04.263022 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:04 crc kubenswrapper[4735]: E1122 08:05:04.263134 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:05:04 crc kubenswrapper[4735]: E1122 08:05:04.264122 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:05:04 crc kubenswrapper[4735]: I1122 08:05:04.264421 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:04 crc kubenswrapper[4735]: E1122 08:05:04.264656 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:05:04 crc kubenswrapper[4735]: E1122 08:05:04.352868 4735 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:05:05 crc kubenswrapper[4735]: I1122 08:05:05.262821 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:05 crc kubenswrapper[4735]: E1122 08:05:05.263012 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:05:06 crc kubenswrapper[4735]: I1122 08:05:06.263183 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:06 crc kubenswrapper[4735]: I1122 08:05:06.263280 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:06 crc kubenswrapper[4735]: E1122 08:05:06.263358 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:05:06 crc kubenswrapper[4735]: I1122 08:05:06.263291 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:06 crc kubenswrapper[4735]: E1122 08:05:06.263585 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:05:06 crc kubenswrapper[4735]: E1122 08:05:06.263604 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:05:07 crc kubenswrapper[4735]: I1122 08:05:07.263290 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:07 crc kubenswrapper[4735]: E1122 08:05:07.263708 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:05:08 crc kubenswrapper[4735]: I1122 08:05:08.262427 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:08 crc kubenswrapper[4735]: E1122 08:05:08.262557 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:05:08 crc kubenswrapper[4735]: I1122 08:05:08.262703 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:08 crc kubenswrapper[4735]: I1122 08:05:08.262896 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:08 crc kubenswrapper[4735]: E1122 08:05:08.263038 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:05:08 crc kubenswrapper[4735]: E1122 08:05:08.263351 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:05:08 crc kubenswrapper[4735]: I1122 08:05:08.263393 4735 scope.go:117] "RemoveContainer" containerID="c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9" Nov 22 08:05:09 crc kubenswrapper[4735]: I1122 08:05:09.012931 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/3.log" Nov 22 08:05:09 crc kubenswrapper[4735]: I1122 08:05:09.016264 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerStarted","Data":"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848"} Nov 22 08:05:09 crc kubenswrapper[4735]: I1122 08:05:09.016642 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:05:09 crc kubenswrapper[4735]: I1122 08:05:09.051842 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podStartSLOduration=109.051829556 podStartE2EDuration="1m49.051829556s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:09.049868423 +0000 UTC m=+130.654207028" watchObservedRunningTime="2025-11-22 08:05:09.051829556 +0000 UTC m=+130.656168161" Nov 22 08:05:09 crc kubenswrapper[4735]: I1122 08:05:09.234522 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vvf57"] Nov 22 08:05:09 crc kubenswrapper[4735]: I1122 08:05:09.234625 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:09 crc kubenswrapper[4735]: E1122 08:05:09.234739 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:05:09 crc kubenswrapper[4735]: E1122 08:05:09.353549 4735 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:05:10 crc kubenswrapper[4735]: I1122 08:05:10.263203 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:10 crc kubenswrapper[4735]: I1122 08:05:10.263290 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:10 crc kubenswrapper[4735]: I1122 08:05:10.263346 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:10 crc kubenswrapper[4735]: E1122 08:05:10.263450 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:05:10 crc kubenswrapper[4735]: E1122 08:05:10.263616 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:05:10 crc kubenswrapper[4735]: E1122 08:05:10.263839 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:05:10 crc kubenswrapper[4735]: I1122 08:05:10.264088 4735 scope.go:117] "RemoveContainer" containerID="a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73" Nov 22 08:05:11 crc kubenswrapper[4735]: I1122 08:05:11.025912 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/1.log" Nov 22 08:05:11 crc kubenswrapper[4735]: I1122 08:05:11.026294 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-85z2h" event={"ID":"61dbab7f-c903-4b55-9d90-aacfd12f7f0f","Type":"ContainerStarted","Data":"8f71d0c8bea2f1c5a42c8365198e3fe40631e28bfc7c5f174a032e124a354f1a"} Nov 22 08:05:11 crc kubenswrapper[4735]: I1122 08:05:11.262794 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:11 crc kubenswrapper[4735]: E1122 08:05:11.262954 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:05:12 crc kubenswrapper[4735]: I1122 08:05:12.262559 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:12 crc kubenswrapper[4735]: I1122 08:05:12.262558 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:12 crc kubenswrapper[4735]: E1122 08:05:12.262701 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:05:12 crc kubenswrapper[4735]: I1122 08:05:12.262587 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:12 crc kubenswrapper[4735]: E1122 08:05:12.262875 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:05:12 crc kubenswrapper[4735]: E1122 08:05:12.263054 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:05:13 crc kubenswrapper[4735]: I1122 08:05:13.263428 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:13 crc kubenswrapper[4735]: E1122 08:05:13.263699 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vvf57" podUID="e9170f64-5b0d-4b2b-99ce-7ecb9f567620" Nov 22 08:05:14 crc kubenswrapper[4735]: I1122 08:05:14.262827 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:14 crc kubenswrapper[4735]: I1122 08:05:14.262854 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:14 crc kubenswrapper[4735]: I1122 08:05:14.262921 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:14 crc kubenswrapper[4735]: E1122 08:05:14.263085 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 08:05:14 crc kubenswrapper[4735]: E1122 08:05:14.263377 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 08:05:14 crc kubenswrapper[4735]: E1122 08:05:14.263683 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 08:05:15 crc kubenswrapper[4735]: I1122 08:05:15.263363 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:15 crc kubenswrapper[4735]: I1122 08:05:15.266003 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 22 08:05:15 crc kubenswrapper[4735]: I1122 08:05:15.266453 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 22 08:05:16 crc kubenswrapper[4735]: I1122 08:05:16.262402 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:16 crc kubenswrapper[4735]: I1122 08:05:16.262487 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:16 crc kubenswrapper[4735]: I1122 08:05:16.262452 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:16 crc kubenswrapper[4735]: I1122 08:05:16.266283 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 22 08:05:16 crc kubenswrapper[4735]: I1122 08:05:16.266574 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 22 08:05:16 crc kubenswrapper[4735]: I1122 08:05:16.267160 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 22 08:05:16 crc kubenswrapper[4735]: I1122 08:05:16.267379 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.311429 4735 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.348398 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-w972q"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.349151 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.354862 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rxw97"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.358003 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.358309 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.358496 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.358389 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.359020 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.359218 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.363679 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.364353 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.364569 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.364372 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-98dz8"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.365901 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.366140 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pc9bj"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.366791 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.372621 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.373237 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.373403 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.373810 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.373903 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.374163 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.380997 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.381302 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.381309 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.381582 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.381626 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382198 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382256 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/994a0303-2bc7-4487-9b70-2733fe287a39-node-pullsecrets\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382294 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-etcd-serving-ca\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382329 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1f800432-2b0d-4452-902e-36d90639a132-audit-dir\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382361 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-encryption-config\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382395 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382427 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-machine-approver-tls\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382482 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-etcd-client\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382514 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpn6w\" (UniqueName: \"kubernetes.io/projected/994a0303-2bc7-4487-9b70-2733fe287a39-kube-api-access-mpn6w\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382545 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/510ea1b6-d5f7-42eb-b490-6f729076dcc3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382619 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382671 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e64c4339-943c-48a4-9809-ad87c03cd233-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wdh9p\" (UID: \"e64c4339-943c-48a4-9809-ad87c03cd233\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382731 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382770 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-image-import-ca\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382800 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-audit\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382823 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-config\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382841 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382862 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-serving-cert\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382903 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47kt7\" (UniqueName: \"kubernetes.io/projected/e64c4339-943c-48a4-9809-ad87c03cd233-kube-api-access-47kt7\") pod \"cluster-samples-operator-665b6dd947-wdh9p\" (UID: \"e64c4339-943c-48a4-9809-ad87c03cd233\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.382933 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-auth-proxy-config\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383040 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383083 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383139 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/510ea1b6-d5f7-42eb-b490-6f729076dcc3-images\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383172 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sggx4\" (UniqueName: \"kubernetes.io/projected/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-kube-api-access-sggx4\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383206 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-config\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383239 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-audit-policies\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383274 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/510ea1b6-d5f7-42eb-b490-6f729076dcc3-config\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383302 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994a0303-2bc7-4487-9b70-2733fe287a39-audit-dir\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383336 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383371 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383404 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383494 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5m5p\" (UniqueName: \"kubernetes.io/projected/510ea1b6-d5f7-42eb-b490-6f729076dcc3-kube-api-access-d5m5p\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383530 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383533 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383710 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.383781 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfxk\" (UniqueName: \"kubernetes.io/projected/1f800432-2b0d-4452-902e-36d90639a132-kube-api-access-8dfxk\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.384063 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.384295 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.384613 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.384670 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6c285"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.385290 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.385413 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.386316 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.386969 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.386983 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.387228 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.387504 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.387734 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.387929 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.388166 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.388588 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.388869 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.389070 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.391737 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.392081 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.392315 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.392535 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.392808 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.392850 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.398680 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.398985 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.400188 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.400628 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.400817 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.425481 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zqccv"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.435132 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.437608 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.440005 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.442740 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.452670 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.453128 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.453217 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.453343 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.453451 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.453499 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.453595 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.453390 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.453397 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.455865 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.455956 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-sjvm9"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.456259 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mtqwx"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.456683 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-l69gd"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.456926 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.456953 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.457035 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.457133 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.457159 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-l69gd" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.457955 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.458237 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.458305 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.458391 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.465341 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.468668 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.469117 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.469265 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.469587 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.470050 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.470188 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.470527 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rv5w5"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.470822 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.470946 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.470905 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.470854 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.471490 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9r74q"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.471809 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8vp7b"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.472119 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.473785 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.490266 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.493503 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.493875 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.494237 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.498092 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.498192 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.498097 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.510518 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.510565 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.510872 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.510885 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.511112 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.511123 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.511234 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.511731 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.512396 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.512625 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.512806 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.512880 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.512943 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513007 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513073 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513084 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513139 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513203 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513276 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513446 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7w2f5"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513498 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513673 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513817 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513839 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513847 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513920 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513943 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513977 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514018 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514022 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514058 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513943 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.513943 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514171 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514192 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514274 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514175 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514607 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514634 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514651 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764f8575-73dd-4450-aefe-0890671a8f98-config\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514677 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/764f8575-73dd-4450-aefe-0890671a8f98-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514699 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-config\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514711 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514723 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfxk\" (UniqueName: \"kubernetes.io/projected/1f800432-2b0d-4452-902e-36d90639a132-kube-api-access-8dfxk\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514750 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-client\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514770 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514812 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514833 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/994a0303-2bc7-4487-9b70-2733fe287a39-node-pullsecrets\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514854 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-etcd-serving-ca\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.514996 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515310 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/994a0303-2bc7-4487-9b70-2733fe287a39-node-pullsecrets\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515534 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7k7n"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515633 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515708 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515730 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-encryption-config\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515769 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twj7q\" (UniqueName: \"kubernetes.io/projected/048ff883-805e-4a2e-b939-86de1aa2e6e2-kube-api-access-twj7q\") pod \"downloads-7954f5f757-l69gd\" (UID: \"048ff883-805e-4a2e-b939-86de1aa2e6e2\") " pod="openshift-console/downloads-7954f5f757-l69gd" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515797 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1f800432-2b0d-4452-902e-36d90639a132-audit-dir\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515825 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-encryption-config\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515850 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpn6w\" (UniqueName: \"kubernetes.io/projected/994a0303-2bc7-4487-9b70-2733fe287a39-kube-api-access-mpn6w\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515873 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d460a3fd-a303-4c45-85ba-23241d0e8976-serving-cert\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515920 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515931 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-serving-cert\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515951 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1f800432-2b0d-4452-902e-36d90639a132-audit-dir\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515961 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515922 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-etcd-serving-ca\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.515988 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-machine-approver-tls\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.516017 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-etcd-client\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.516038 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/510ea1b6-d5f7-42eb-b490-6f729076dcc3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.516056 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.516386 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.516724 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.516983 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.516073 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e64c4339-943c-48a4-9809-ad87c03cd233-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wdh9p\" (UID: \"e64c4339-943c-48a4-9809-ad87c03cd233\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517353 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517370 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-image-import-ca\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517385 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-audit\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517406 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qddmw\" (UniqueName: \"kubernetes.io/projected/d460a3fd-a303-4c45-85ba-23241d0e8976-kube-api-access-qddmw\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517421 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-audit-policies\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517439 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvkkl\" (UniqueName: \"kubernetes.io/projected/a9b8f278-7174-4b90-b14f-6091cacdb5c2-kube-api-access-fvkkl\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517475 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-config\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517491 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517510 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-service-ca\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517527 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-serving-cert\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517543 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9b8f278-7174-4b90-b14f-6091cacdb5c2-audit-dir\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517559 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47kt7\" (UniqueName: \"kubernetes.io/projected/e64c4339-943c-48a4-9809-ad87c03cd233-kube-api-access-47kt7\") pod \"cluster-samples-operator-665b6dd947-wdh9p\" (UID: \"e64c4339-943c-48a4-9809-ad87c03cd233\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517574 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517606 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-auth-proxy-config\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517622 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/764f8575-73dd-4450-aefe-0890671a8f98-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517638 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnpbm\" (UniqueName: \"kubernetes.io/projected/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-kube-api-access-bnpbm\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517658 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-ca\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517679 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adbace7e-3d6e-43d5-9331-af5f0b57fe20-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517696 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517716 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517733 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517941 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-serving-cert\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517960 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9246bb69-d48c-40bb-bb3f-1a02872bf250-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517965 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.517991 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/510ea1b6-d5f7-42eb-b490-6f729076dcc3-images\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518010 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/adbace7e-3d6e-43d5-9331-af5f0b57fe20-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518028 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9246bb69-d48c-40bb-bb3f-1a02872bf250-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518044 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9q7s\" (UniqueName: \"kubernetes.io/projected/adbace7e-3d6e-43d5-9331-af5f0b57fe20-kube-api-access-f9q7s\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518060 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxxdb\" (UniqueName: \"kubernetes.io/projected/9246bb69-d48c-40bb-bb3f-1a02872bf250-kube-api-access-hxxdb\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518080 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sggx4\" (UniqueName: \"kubernetes.io/projected/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-kube-api-access-sggx4\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518102 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-config\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518120 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adbace7e-3d6e-43d5-9331-af5f0b57fe20-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518137 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-audit-policies\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518152 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994a0303-2bc7-4487-9b70-2733fe287a39-audit-dir\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518169 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-etcd-client\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518188 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/510ea1b6-d5f7-42eb-b490-6f729076dcc3-config\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518207 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518227 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518244 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518265 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5m5p\" (UniqueName: \"kubernetes.io/projected/510ea1b6-d5f7-42eb-b490-6f729076dcc3-kube-api-access-d5m5p\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518281 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518299 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518448 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-audit\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.518792 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.519171 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-config\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.519245 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.519514 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/510ea1b6-d5f7-42eb-b490-6f729076dcc3-images\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.519768 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/994a0303-2bc7-4487-9b70-2733fe287a39-image-import-ca\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.519789 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/510ea1b6-d5f7-42eb-b490-6f729076dcc3-config\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.519851 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994a0303-2bc7-4487-9b70-2733fe287a39-audit-dir\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.520109 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.520134 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.520451 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.520875 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-auth-proxy-config\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.522597 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xksr2"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.523009 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.523211 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.523417 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.523583 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.523783 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.523929 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-etcd-client\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.525291 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-audit-policies\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.527255 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.529734 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.529834 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.533219 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.534695 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.535712 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.536122 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-encryption-config\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.536428 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e64c4339-943c-48a4-9809-ad87c03cd233-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wdh9p\" (UID: \"e64c4339-943c-48a4-9809-ad87c03cd233\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.536637 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/510ea1b6-d5f7-42eb-b490-6f729076dcc3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.537395 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.537983 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zm9cl"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.538295 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.538327 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.538608 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.540800 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-config\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.540902 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.545555 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.546301 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.549014 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.549098 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.549710 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-w5v4q"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.550098 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rxw97"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.550187 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.550675 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.554325 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6c285"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.563829 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.561114 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.561234 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.563770 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-machine-approver-tls\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.593034 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.593169 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/994a0303-2bc7-4487-9b70-2733fe287a39-serving-cert\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.594198 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.594528 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.595929 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.594585 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.596008 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.597872 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.597899 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mtqwx"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.597908 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.597908 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.597918 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-98dz8"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.597986 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.601007 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.608803 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.611341 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.614647 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pc9bj"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.617672 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-l69gd"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.618979 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/764f8575-73dd-4450-aefe-0890671a8f98-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619019 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnpbm\" (UniqueName: \"kubernetes.io/projected/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-kube-api-access-bnpbm\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619046 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-ca\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619069 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adbace7e-3d6e-43d5-9331-af5f0b57fe20-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619089 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619108 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-serving-cert\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619128 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9246bb69-d48c-40bb-bb3f-1a02872bf250-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619162 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/adbace7e-3d6e-43d5-9331-af5f0b57fe20-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619185 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9246bb69-d48c-40bb-bb3f-1a02872bf250-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619216 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9q7s\" (UniqueName: \"kubernetes.io/projected/adbace7e-3d6e-43d5-9331-af5f0b57fe20-kube-api-access-f9q7s\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619239 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxxdb\" (UniqueName: \"kubernetes.io/projected/9246bb69-d48c-40bb-bb3f-1a02872bf250-kube-api-access-hxxdb\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619260 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adbace7e-3d6e-43d5-9331-af5f0b57fe20-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619285 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-etcd-client\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619370 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764f8575-73dd-4450-aefe-0890671a8f98-config\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619398 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/764f8575-73dd-4450-aefe-0890671a8f98-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619444 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-config\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619482 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-client\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619505 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619545 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-encryption-config\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619568 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twj7q\" (UniqueName: \"kubernetes.io/projected/048ff883-805e-4a2e-b939-86de1aa2e6e2-kube-api-access-twj7q\") pod \"downloads-7954f5f757-l69gd\" (UID: \"048ff883-805e-4a2e-b939-86de1aa2e6e2\") " pod="openshift-console/downloads-7954f5f757-l69gd" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619590 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-serving-cert\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619621 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d460a3fd-a303-4c45-85ba-23241d0e8976-serving-cert\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619646 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qddmw\" (UniqueName: \"kubernetes.io/projected/d460a3fd-a303-4c45-85ba-23241d0e8976-kube-api-access-qddmw\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619667 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-audit-policies\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619692 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-service-ca\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619713 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvkkl\" (UniqueName: \"kubernetes.io/projected/a9b8f278-7174-4b90-b14f-6091cacdb5c2-kube-api-access-fvkkl\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619735 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9b8f278-7174-4b90-b14f-6091cacdb5c2-audit-dir\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.619770 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.620255 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.620674 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-config\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.621786 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-ca\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.622057 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9b8f278-7174-4b90-b14f-6091cacdb5c2-audit-dir\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.622327 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-service-ca\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.622821 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764f8575-73dd-4450-aefe-0890671a8f98-config\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.622865 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9246bb69-d48c-40bb-bb3f-1a02872bf250-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.622905 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-audit-policies\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.623093 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zqccv"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.623239 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adbace7e-3d6e-43d5-9331-af5f0b57fe20-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.623377 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a9b8f278-7174-4b90-b14f-6091cacdb5c2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.623640 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-serving-cert\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.625201 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d460a3fd-a303-4c45-85ba-23241d0e8976-etcd-client\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.625568 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-encryption-config\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.625613 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.626085 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d460a3fd-a303-4c45-85ba-23241d0e8976-serving-cert\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.626941 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9246bb69-d48c-40bb-bb3f-1a02872bf250-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.627682 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9r74q"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.629480 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.630163 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/adbace7e-3d6e-43d5-9331-af5f0b57fe20-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.630566 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a9b8f278-7174-4b90-b14f-6091cacdb5c2-etcd-client\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.630844 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.631427 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.633052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.633893 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4t74q"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.634935 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.635897 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-serving-cert\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.636056 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.637289 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sjvm9"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.638264 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.638446 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/764f8575-73dd-4450-aefe-0890671a8f98-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.640180 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8vp7b"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.641184 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7k7n"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.642166 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.643322 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.644078 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.645091 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.646233 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rv5w5"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.647775 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.648262 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7w2f5"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.649214 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xksr2"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.650159 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-wrfpv"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.651089 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.651088 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-5s2w9"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.652113 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.652130 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.652719 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.653692 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4t74q"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.654705 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.656966 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.658041 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zm9cl"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.659186 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.660625 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.661646 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.662640 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5s2w9"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.663629 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-xkl5g"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.664170 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xkl5g" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.664893 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xkl5g"] Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.671737 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.691054 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.711883 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.731607 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.751875 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.771652 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.791401 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.811905 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.830884 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.851705 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.879344 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.892011 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.911321 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.931757 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.959591 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.971154 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 22 08:05:19 crc kubenswrapper[4735]: I1122 08:05:19.991787 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.011296 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.032590 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.051703 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.071172 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.091618 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.111590 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.131777 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.151028 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.172667 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.193422 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.211945 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.232631 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.252067 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.292920 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.301256 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfxk\" (UniqueName: \"kubernetes.io/projected/1f800432-2b0d-4452-902e-36d90639a132-kube-api-access-8dfxk\") pod \"oauth-openshift-558db77b4-pc9bj\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.312720 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.332342 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.351524 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.392280 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.396683 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpn6w\" (UniqueName: \"kubernetes.io/projected/994a0303-2bc7-4487-9b70-2733fe287a39-kube-api-access-mpn6w\") pod \"apiserver-76f77b778f-rxw97\" (UID: \"994a0303-2bc7-4487-9b70-2733fe287a39\") " pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.411170 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.421033 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.440914 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.453651 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.491343 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47kt7\" (UniqueName: \"kubernetes.io/projected/e64c4339-943c-48a4-9809-ad87c03cd233-kube-api-access-47kt7\") pod \"cluster-samples-operator-665b6dd947-wdh9p\" (UID: \"e64c4339-943c-48a4-9809-ad87c03cd233\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.512317 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sggx4\" (UniqueName: \"kubernetes.io/projected/bcc3f64f-86dd-459c-8921-5cd3cf704ea5-kube-api-access-sggx4\") pod \"machine-approver-56656f9798-w972q\" (UID: \"bcc3f64f-86dd-459c-8921-5cd3cf704ea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.530394 4735 request.go:700] Waited for 1.009959253s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.532690 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.535697 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5m5p\" (UniqueName: \"kubernetes.io/projected/510ea1b6-d5f7-42eb-b490-6f729076dcc3-kube-api-access-d5m5p\") pod \"machine-api-operator-5694c8668f-98dz8\" (UID: \"510ea1b6-d5f7-42eb-b490-6f729076dcc3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.577705 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.578023 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.599787 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631046 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ddc6014-cf08-45d5-8cc9-56fdfd119657-metrics-tls\") pod \"dns-operator-744455d44c-mtqwx\" (UID: \"5ddc6014-cf08-45d5-8cc9-56fdfd119657\") " pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631093 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d841f5e-d73a-4777-acc1-7b28f247e829-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631121 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktgps\" (UniqueName: \"kubernetes.io/projected/5b742605-ede0-46c6-add0-01cda9fa184d-kube-api-access-ktgps\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631151 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-console-config\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631177 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-service-ca-bundle\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631200 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-config\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631225 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-oauth-config\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631248 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-tls\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631302 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631343 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7hhj\" (UniqueName: \"kubernetes.io/projected/5ddc6014-cf08-45d5-8cc9-56fdfd119657-kube-api-access-d7hhj\") pod \"dns-operator-744455d44c-mtqwx\" (UID: \"5ddc6014-cf08-45d5-8cc9-56fdfd119657\") " pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631365 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631418 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b742605-ede0-46c6-add0-01cda9fa184d-serving-cert\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631471 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e095b65e-6f13-4a89-8418-78f5bda82cb5-serving-cert\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.631490 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.632481 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-config\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.632551 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zlj6\" (UniqueName: \"kubernetes.io/projected/2d841f5e-d73a-4777-acc1-7b28f247e829-kube-api-access-4zlj6\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.632611 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjh2j\" (UniqueName: \"kubernetes.io/projected/5dea56bb-e794-479f-8269-e31474c18920-kube-api-access-tjh2j\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.632649 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-oauth-serving-cert\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.632714 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-certificates\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.632739 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d841f5e-d73a-4777-acc1-7b28f247e829-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.632776 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dea56bb-e794-479f-8269-e31474c18920-config\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.632848 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h58m\" (UniqueName: \"kubernetes.io/projected/d417a7ee-7687-4086-8047-39e259a9bfbe-kube-api-access-6h58m\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633590 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dea56bb-e794-479f-8269-e31474c18920-trusted-ca\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633671 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-bound-sa-token\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633725 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlfq7\" (UniqueName: \"kubernetes.io/projected/e095b65e-6f13-4a89-8418-78f5bda82cb5-kube-api-access-dlfq7\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633779 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf2hq\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-kube-api-access-cf2hq\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633801 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-serving-cert\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633858 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-client-ca\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633887 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-service-ca\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633933 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.633957 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5dea56bb-e794-479f-8269-e31474c18920-serving-cert\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.634017 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.634043 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-trusted-ca\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.634087 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-trusted-ca-bundle\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.634390 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: E1122 08:05:20.635769 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.135756252 +0000 UTC m=+142.740094857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.651821 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.653453 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.671621 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.681833 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pc9bj"] Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.691413 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 22 08:05:20 crc kubenswrapper[4735]: W1122 08:05:20.691821 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f800432_2b0d_4452_902e_36d90639a132.slice/crio-d76588080016ba1d1ff8a49b9b7765c8c0a036161089a7b622af522a5b832fde WatchSource:0}: Error finding container d76588080016ba1d1ff8a49b9b7765c8c0a036161089a7b622af522a5b832fde: Status 404 returned error can't find the container with id d76588080016ba1d1ff8a49b9b7765c8c0a036161089a7b622af522a5b832fde Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.711345 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.711873 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.712212 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.735393 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.735801 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:20 crc kubenswrapper[4735]: E1122 08:05:20.736134 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.235895493 +0000 UTC m=+142.840234098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736160 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-client-ca\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736205 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57632601-15a1-4aed-a66b-3f01331cb637-proxy-tls\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736231 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktgps\" (UniqueName: \"kubernetes.io/projected/5b742605-ede0-46c6-add0-01cda9fa184d-kube-api-access-ktgps\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736254 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e8094c96-1edb-431d-8081-4a4927be57a8-tmpfs\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736277 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-socket-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736317 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d841f5e-d73a-4777-acc1-7b28f247e829-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736356 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-console-config\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736381 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xhkl\" (UniqueName: \"kubernetes.io/projected/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-kube-api-access-5xhkl\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736403 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736427 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-service-ca-bundle\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736450 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e9f768a-e958-41b3-975e-45f565437df6-cert\") pod \"ingress-canary-xkl5g\" (UID: \"0e9f768a-e958-41b3-975e-45f565437df6\") " pod="openshift-ingress-canary/ingress-canary-xkl5g" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736521 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38819d0b-58b8-429b-9969-5309efa8d187-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736547 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kwhz\" (UniqueName: \"kubernetes.io/projected/fb194fbb-af18-420c-ad19-004eba55e639-kube-api-access-7kwhz\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzwms\" (UID: \"fb194fbb-af18-420c-ad19-004eba55e639\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736590 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-oauth-config\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736611 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2cf18025-0832-4850-8568-9af2a131711c-metrics-tls\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736663 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7hhj\" (UniqueName: \"kubernetes.io/projected/5ddc6014-cf08-45d5-8cc9-56fdfd119657-kube-api-access-d7hhj\") pod \"dns-operator-744455d44c-mtqwx\" (UID: \"5ddc6014-cf08-45d5-8cc9-56fdfd119657\") " pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736687 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736725 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4fvp\" (UniqueName: \"kubernetes.io/projected/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-kube-api-access-w4fvp\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736763 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bf91eba7-b4d7-486b-98e1-2ca53735adc5-signing-key\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736788 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03719f40-3ec6-416a-ba6e-33e7ebd963c6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736812 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c22gx\" (UniqueName: \"kubernetes.io/projected/38819d0b-58b8-429b-9969-5309efa8d187-kube-api-access-c22gx\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736837 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/674120ce-7924-489f-ab5c-f39b39df718f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736861 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-serving-cert\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736910 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/674120ce-7924-489f-ab5c-f39b39df718f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.736933 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a47af7c4-72b1-45bc-95c4-5f3207307fa2-trusted-ca\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738652 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zlj6\" (UniqueName: \"kubernetes.io/projected/2d841f5e-d73a-4777-acc1-7b28f247e829-kube-api-access-4zlj6\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738681 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjh2j\" (UniqueName: \"kubernetes.io/projected/5dea56bb-e794-479f-8269-e31474c18920-kube-api-access-tjh2j\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738706 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-profile-collector-cert\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738736 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlg27\" (UniqueName: \"kubernetes.io/projected/9f713337-be7d-47e8-bc15-76b5795d3c43-kube-api-access-jlg27\") pod \"migrator-59844c95c7-dl4cp\" (UID: \"9f713337-be7d-47e8-bc15-76b5795d3c43\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738762 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03719f40-3ec6-416a-ba6e-33e7ebd963c6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738829 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-certificates\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738832 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738880 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb5ct\" (UniqueName: \"kubernetes.io/projected/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-kube-api-access-fb5ct\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738906 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a47af7c4-72b1-45bc-95c4-5f3207307fa2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738945 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-srv-cert\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738977 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h58m\" (UniqueName: \"kubernetes.io/projected/d417a7ee-7687-4086-8047-39e259a9bfbe-kube-api-access-6h58m\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738607 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-console-config\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.739001 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1296f167-e81f-4641-b1b0-d0de2b031d63-certs\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.737548 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d841f5e-d73a-4777-acc1-7b28f247e829-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.738450 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-service-ca-bundle\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.739282 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dea56bb-e794-479f-8269-e31474c18920-trusted-ca\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.739312 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-config\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.739333 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.739359 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-config\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740340 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-bound-sa-token\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740383 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlfq7\" (UniqueName: \"kubernetes.io/projected/e095b65e-6f13-4a89-8418-78f5bda82cb5-kube-api-access-dlfq7\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740404 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-serving-cert\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740441 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-client-ca\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740527 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7dn\" (UniqueName: \"kubernetes.io/projected/a1f75b9e-1a13-4f13-8a78-98111d8dfeaf-kube-api-access-wf7dn\") pod \"package-server-manager-789f6589d5-g5sst\" (UID: \"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740566 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dea56bb-e794-479f-8269-e31474c18920-trusted-ca\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740682 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e8094c96-1edb-431d-8081-4a4927be57a8-apiservice-cert\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740724 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a47af7c4-72b1-45bc-95c4-5f3207307fa2-metrics-tls\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740751 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740792 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bf91eba7-b4d7-486b-98e1-2ca53735adc5-signing-cabundle\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740912 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-default-certificate\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.740997 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e8094c96-1edb-431d-8081-4a4927be57a8-webhook-cert\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.741123 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-trusted-ca-bundle\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.741291 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-client-ca\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.741301 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.741360 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-trusted-ca\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: E1122 08:05:20.741586 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.241570778 +0000 UTC m=+142.845909453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.741585 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-certificates\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.741970 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc9cf\" (UniqueName: \"kubernetes.io/projected/31a557ae-6002-49fd-a9af-b3ea23dfd952-kube-api-access-sc9cf\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.741998 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-oauth-config\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742259 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-stats-auth\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742283 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-trusted-ca-bundle\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742300 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ddc6014-cf08-45d5-8cc9-56fdfd119657-metrics-tls\") pod \"dns-operator-744455d44c-mtqwx\" (UID: \"5ddc6014-cf08-45d5-8cc9-56fdfd119657\") " pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742327 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38819d0b-58b8-429b-9969-5309efa8d187-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742352 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742376 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjdtq\" (UniqueName: \"kubernetes.io/projected/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-kube-api-access-xjdtq\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742398 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6l48\" (UniqueName: \"kubernetes.io/projected/2cf18025-0832-4850-8568-9af2a131711c-kube-api-access-q6l48\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742444 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2da8d731-f76e-472f-bb05-b2242fb9a652-serving-cert\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742479 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-trusted-ca\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742494 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjlcg\" (UniqueName: \"kubernetes.io/projected/53cd51d1-1576-492a-ba5e-e26a388b6a57-kube-api-access-cjlcg\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742820 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-registration-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742871 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-config\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.742969 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-tls\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743017 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743042 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5r5w\" (UniqueName: \"kubernetes.io/projected/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-kube-api-access-s5r5w\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743075 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b742605-ede0-46c6-add0-01cda9fa184d-serving-cert\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743098 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1296f167-e81f-4641-b1b0-d0de2b031d63-node-bootstrap-token\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743152 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e095b65e-6f13-4a89-8418-78f5bda82cb5-serving-cert\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743245 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743274 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-proxy-tls\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743307 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl62v\" (UniqueName: \"kubernetes.io/projected/1296f167-e81f-4641-b1b0-d0de2b031d63-kube-api-access-hl62v\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743334 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-config\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743508 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-metrics-certs\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743541 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-mountpoint-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743567 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdhrz\" (UniqueName: \"kubernetes.io/projected/a47af7c4-72b1-45bc-95c4-5f3207307fa2-kube-api-access-fdhrz\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743608 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-images\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743632 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6hhs\" (UniqueName: \"kubernetes.io/projected/2da8d731-f76e-472f-bb05-b2242fb9a652-kube-api-access-n6hhs\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743719 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-oauth-serving-cert\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743751 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n5z9\" (UniqueName: \"kubernetes.io/projected/57632601-15a1-4aed-a66b-3f01331cb637-kube-api-access-5n5z9\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743814 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d841f5e-d73a-4777-acc1-7b28f247e829-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743843 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpcmx\" (UniqueName: \"kubernetes.io/projected/0e9f768a-e958-41b3-975e-45f565437df6-kube-api-access-lpcmx\") pod \"ingress-canary-xkl5g\" (UID: \"0e9f768a-e958-41b3-975e-45f565437df6\") " pod="openshift-ingress-canary/ingress-canary-xkl5g" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743872 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dea56bb-e794-479f-8269-e31474c18920-config\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743895 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4b91e784-c670-41af-95c6-696a7197e01a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7w2f5\" (UID: \"4b91e784-c670-41af-95c6-696a7197e01a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743922 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.743956 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-plugins-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.744008 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57632601-15a1-4aed-a66b-3f01331cb637-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.744030 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82nm9\" (UniqueName: \"kubernetes.io/projected/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-kube-api-access-82nm9\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.744052 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfzhx\" (UniqueName: \"kubernetes.io/projected/bf91eba7-b4d7-486b-98e1-2ca53735adc5-kube-api-access-tfzhx\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.744077 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f75b9e-1a13-4f13-8a78-98111d8dfeaf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g5sst\" (UID: \"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.744118 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf2hq\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-kube-api-access-cf2hq\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.746734 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-config\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.746937 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-service-ca\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.746978 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5dea56bb-e794-479f-8269-e31474c18920-serving-cert\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.747754 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b742605-ede0-46c6-add0-01cda9fa184d-config\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.747874 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/674120ce-7924-489f-ab5c-f39b39df718f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748101 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2cf18025-0832-4850-8568-9af2a131711c-config-volume\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748152 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqr46\" (UniqueName: \"kubernetes.io/projected/e8094c96-1edb-431d-8081-4a4927be57a8-kube-api-access-fqr46\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748248 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748348 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-service-ca\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748567 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31a557ae-6002-49fd-a9af-b3ea23dfd952-secret-volume\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748602 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb194fbb-af18-420c-ad19-004eba55e639-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzwms\" (UID: \"fb194fbb-af18-420c-ad19-004eba55e639\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748635 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53cd51d1-1576-492a-ba5e-e26a388b6a57-service-ca-bundle\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748715 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31a557ae-6002-49fd-a9af-b3ea23dfd952-config-volume\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748742 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03719f40-3ec6-416a-ba6e-33e7ebd963c6-config\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748827 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dea56bb-e794-479f-8269-e31474c18920-config\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.748938 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7zlw\" (UniqueName: \"kubernetes.io/projected/4b91e784-c670-41af-95c6-696a7197e01a-kube-api-access-n7zlw\") pod \"multus-admission-controller-857f4d67dd-7w2f5\" (UID: \"4b91e784-c670-41af-95c6-696a7197e01a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.749005 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-srv-cert\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.749036 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-csi-data-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.749040 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-oauth-serving-cert\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.749362 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.751769 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e095b65e-6f13-4a89-8418-78f5bda82cb5-serving-cert\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.751972 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-serving-cert\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.752527 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.752887 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.753269 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-tls\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.753538 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d841f5e-d73a-4777-acc1-7b28f247e829-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.753928 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ddc6014-cf08-45d5-8cc9-56fdfd119657-metrics-tls\") pod \"dns-operator-744455d44c-mtqwx\" (UID: \"5ddc6014-cf08-45d5-8cc9-56fdfd119657\") " pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.755085 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5dea56bb-e794-479f-8269-e31474c18920-serving-cert\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.756423 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b742605-ede0-46c6-add0-01cda9fa184d-serving-cert\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.772366 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.793013 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.811660 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.829261 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rxw97"] Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.831532 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850004 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:20 crc kubenswrapper[4735]: E1122 08:05:20.850111 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.350087647 +0000 UTC m=+142.954426252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850185 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4fvp\" (UniqueName: \"kubernetes.io/projected/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-kube-api-access-w4fvp\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850217 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bf91eba7-b4d7-486b-98e1-2ca53735adc5-signing-key\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850244 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03719f40-3ec6-416a-ba6e-33e7ebd963c6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850268 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c22gx\" (UniqueName: \"kubernetes.io/projected/38819d0b-58b8-429b-9969-5309efa8d187-kube-api-access-c22gx\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850290 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/674120ce-7924-489f-ab5c-f39b39df718f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850310 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-serving-cert\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850364 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-profile-collector-cert\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850384 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/674120ce-7924-489f-ab5c-f39b39df718f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850406 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a47af7c4-72b1-45bc-95c4-5f3207307fa2-trusted-ca\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850431 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlg27\" (UniqueName: \"kubernetes.io/projected/9f713337-be7d-47e8-bc15-76b5795d3c43-kube-api-access-jlg27\") pod \"migrator-59844c95c7-dl4cp\" (UID: \"9f713337-be7d-47e8-bc15-76b5795d3c43\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850476 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03719f40-3ec6-416a-ba6e-33e7ebd963c6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850510 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb5ct\" (UniqueName: \"kubernetes.io/projected/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-kube-api-access-fb5ct\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850531 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a47af7c4-72b1-45bc-95c4-5f3207307fa2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850550 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-srv-cert\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850590 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1296f167-e81f-4641-b1b0-d0de2b031d63-certs\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-config\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850650 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850672 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-config\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850698 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7dn\" (UniqueName: \"kubernetes.io/projected/a1f75b9e-1a13-4f13-8a78-98111d8dfeaf-kube-api-access-wf7dn\") pod \"package-server-manager-789f6589d5-g5sst\" (UID: \"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.850995 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e8094c96-1edb-431d-8081-4a4927be57a8-apiservice-cert\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851025 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a47af7c4-72b1-45bc-95c4-5f3207307fa2-metrics-tls\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851047 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bf91eba7-b4d7-486b-98e1-2ca53735adc5-signing-cabundle\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851070 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-default-certificate\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851094 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e8094c96-1edb-431d-8081-4a4927be57a8-webhook-cert\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851124 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851256 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc9cf\" (UniqueName: \"kubernetes.io/projected/31a557ae-6002-49fd-a9af-b3ea23dfd952-kube-api-access-sc9cf\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851404 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-stats-auth\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851431 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38819d0b-58b8-429b-9969-5309efa8d187-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851476 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851674 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.851737 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-config\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:20 crc kubenswrapper[4735]: E1122 08:05:20.852172 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.352156184 +0000 UTC m=+142.956494879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.853346 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-config\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.853704 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjdtq\" (UniqueName: \"kubernetes.io/projected/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-kube-api-access-xjdtq\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.853744 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6l48\" (UniqueName: \"kubernetes.io/projected/2cf18025-0832-4850-8568-9af2a131711c-kube-api-access-q6l48\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.853946 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2da8d731-f76e-472f-bb05-b2242fb9a652-serving-cert\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.853975 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjlcg\" (UniqueName: \"kubernetes.io/projected/53cd51d1-1576-492a-ba5e-e26a388b6a57-kube-api-access-cjlcg\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.854108 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bf91eba7-b4d7-486b-98e1-2ca53735adc5-signing-cabundle\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.854201 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-registration-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.854530 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-registration-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.854863 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.854915 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a47af7c4-72b1-45bc-95c4-5f3207307fa2-trusted-ca\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855339 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5r5w\" (UniqueName: \"kubernetes.io/projected/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-kube-api-access-s5r5w\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855406 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1296f167-e81f-4641-b1b0-d0de2b031d63-node-bootstrap-token\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855536 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-srv-cert\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855738 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-proxy-tls\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855772 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl62v\" (UniqueName: \"kubernetes.io/projected/1296f167-e81f-4641-b1b0-d0de2b031d63-kube-api-access-hl62v\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855796 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-metrics-certs\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855821 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-mountpoint-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855863 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdhrz\" (UniqueName: \"kubernetes.io/projected/a47af7c4-72b1-45bc-95c4-5f3207307fa2-kube-api-access-fdhrz\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855887 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-images\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855909 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6hhs\" (UniqueName: \"kubernetes.io/projected/2da8d731-f76e-472f-bb05-b2242fb9a652-kube-api-access-n6hhs\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855943 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n5z9\" (UniqueName: \"kubernetes.io/projected/57632601-15a1-4aed-a66b-3f01331cb637-kube-api-access-5n5z9\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855972 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpcmx\" (UniqueName: \"kubernetes.io/projected/0e9f768a-e958-41b3-975e-45f565437df6-kube-api-access-lpcmx\") pod \"ingress-canary-xkl5g\" (UID: \"0e9f768a-e958-41b3-975e-45f565437df6\") " pod="openshift-ingress-canary/ingress-canary-xkl5g" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.855997 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4b91e784-c670-41af-95c6-696a7197e01a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7w2f5\" (UID: \"4b91e784-c670-41af-95c6-696a7197e01a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856021 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856042 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-plugins-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856057 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e8094c96-1edb-431d-8081-4a4927be57a8-webhook-cert\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856075 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57632601-15a1-4aed-a66b-3f01331cb637-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856105 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82nm9\" (UniqueName: \"kubernetes.io/projected/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-kube-api-access-82nm9\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856269 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfzhx\" (UniqueName: \"kubernetes.io/projected/bf91eba7-b4d7-486b-98e1-2ca53735adc5-kube-api-access-tfzhx\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856296 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f75b9e-1a13-4f13-8a78-98111d8dfeaf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g5sst\" (UID: \"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856348 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqr46\" (UniqueName: \"kubernetes.io/projected/e8094c96-1edb-431d-8081-4a4927be57a8-kube-api-access-fqr46\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856364 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/674120ce-7924-489f-ab5c-f39b39df718f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856366 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-mountpoint-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856382 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2cf18025-0832-4850-8568-9af2a131711c-config-volume\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856405 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31a557ae-6002-49fd-a9af-b3ea23dfd952-secret-volume\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856424 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb194fbb-af18-420c-ad19-004eba55e639-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzwms\" (UID: \"fb194fbb-af18-420c-ad19-004eba55e639\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856593 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-plugins-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856626 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53cd51d1-1576-492a-ba5e-e26a388b6a57-service-ca-bundle\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856696 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31a557ae-6002-49fd-a9af-b3ea23dfd952-config-volume\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856726 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03719f40-3ec6-416a-ba6e-33e7ebd963c6-config\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7zlw\" (UniqueName: \"kubernetes.io/projected/4b91e784-c670-41af-95c6-696a7197e01a-kube-api-access-n7zlw\") pod \"multus-admission-controller-857f4d67dd-7w2f5\" (UID: \"4b91e784-c670-41af-95c6-696a7197e01a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856821 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856848 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-srv-cert\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856875 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-csi-data-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856994 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-csi-data-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.856996 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57632601-15a1-4aed-a66b-3f01331cb637-proxy-tls\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857124 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-client-ca\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857176 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e8094c96-1edb-431d-8081-4a4927be57a8-tmpfs\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857203 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-socket-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857232 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xhkl\" (UniqueName: \"kubernetes.io/projected/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-kube-api-access-5xhkl\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857257 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857286 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e9f768a-e958-41b3-975e-45f565437df6-cert\") pod \"ingress-canary-xkl5g\" (UID: \"0e9f768a-e958-41b3-975e-45f565437df6\") " pod="openshift-ingress-canary/ingress-canary-xkl5g" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857313 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38819d0b-58b8-429b-9969-5309efa8d187-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857339 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kwhz\" (UniqueName: \"kubernetes.io/projected/fb194fbb-af18-420c-ad19-004eba55e639-kube-api-access-7kwhz\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzwms\" (UID: \"fb194fbb-af18-420c-ad19-004eba55e639\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857368 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2cf18025-0832-4850-8568-9af2a131711c-metrics-tls\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.857434 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-socket-dir\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.858043 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03719f40-3ec6-416a-ba6e-33e7ebd963c6-config\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.858259 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-client-ca\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.858858 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31a557ae-6002-49fd-a9af-b3ea23dfd952-config-volume\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.859036 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57632601-15a1-4aed-a66b-3f01331cb637-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.859237 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e8094c96-1edb-431d-8081-4a4927be57a8-apiservice-cert\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.859394 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e8094c96-1edb-431d-8081-4a4927be57a8-tmpfs\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.860520 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4b91e784-c670-41af-95c6-696a7197e01a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7w2f5\" (UID: \"4b91e784-c670-41af-95c6-696a7197e01a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.860968 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03719f40-3ec6-416a-ba6e-33e7ebd963c6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.861034 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2da8d731-f76e-472f-bb05-b2242fb9a652-serving-cert\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.861077 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bf91eba7-b4d7-486b-98e1-2ca53735adc5-signing-key\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.861367 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f75b9e-1a13-4f13-8a78-98111d8dfeaf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g5sst\" (UID: \"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.861591 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb194fbb-af18-420c-ad19-004eba55e639-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzwms\" (UID: \"fb194fbb-af18-420c-ad19-004eba55e639\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.862098 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a47af7c4-72b1-45bc-95c4-5f3207307fa2-metrics-tls\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.862113 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31a557ae-6002-49fd-a9af-b3ea23dfd952-secret-volume\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.863170 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-serving-cert\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.863222 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-profile-collector-cert\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.863708 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.864390 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.865569 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-srv-cert\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.872329 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.890951 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.913167 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.932971 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.946474 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/674120ce-7924-489f-ab5c-f39b39df718f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.951896 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.957327 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/674120ce-7924-489f-ab5c-f39b39df718f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.958507 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:20 crc kubenswrapper[4735]: E1122 08:05:20.958694 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.458645648 +0000 UTC m=+143.062984263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.959153 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:20 crc kubenswrapper[4735]: E1122 08:05:20.959502 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.459493921 +0000 UTC m=+143.063832526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.971801 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 22 08:05:20 crc kubenswrapper[4735]: I1122 08:05:20.991799 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.012039 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.017078 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-default-certificate\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.033403 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.046448 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-stats-auth\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.052136 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.059950 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.060096 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.560074244 +0000 UTC m=+143.164412849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.060610 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.060988 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.560979059 +0000 UTC m=+143.165317664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.062650 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53cd51d1-1576-492a-ba5e-e26a388b6a57-metrics-certs\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.068364 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" event={"ID":"1f800432-2b0d-4452-902e-36d90639a132","Type":"ContainerStarted","Data":"ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67"} Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.068419 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" event={"ID":"1f800432-2b0d-4452-902e-36d90639a132","Type":"ContainerStarted","Data":"d76588080016ba1d1ff8a49b9b7765c8c0a036161089a7b622af522a5b832fde"} Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.069203 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.071129 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" event={"ID":"994a0303-2bc7-4487-9b70-2733fe287a39","Type":"ContainerStarted","Data":"de3fdbfd41be1416fb4a9a1a55b70fec9518418dcedd8b296b53b5cbb8825ca9"} Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.071661 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.072335 4735 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pc9bj container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" start-of-body= Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.072374 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" podUID="1f800432-2b0d-4452-902e-36d90639a132" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.073696 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" event={"ID":"bcc3f64f-86dd-459c-8921-5cd3cf704ea5","Type":"ContainerStarted","Data":"6d425e453061aa94149e04014966180bf6b4537d46452c3663d6ec1c1dd472d7"} Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.073742 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" event={"ID":"bcc3f64f-86dd-459c-8921-5cd3cf704ea5","Type":"ContainerStarted","Data":"86781cf7138b60bc6caa31a35da27bfbf444c0f3142a011d59fd00d37cfcdc05"} Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.091055 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.097605 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53cd51d1-1576-492a-ba5e-e26a388b6a57-service-ca-bundle\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.111676 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.132181 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.138805 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p"] Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.140106 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-98dz8"] Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.154757 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.161602 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.161726 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.661708955 +0000 UTC m=+143.266047560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.161901 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.162245 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38819d0b-58b8-429b-9969-5309efa8d187-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.162296 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.662285882 +0000 UTC m=+143.266624487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: W1122 08:05:21.164973 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod510ea1b6_d5f7_42eb_b490_6f729076dcc3.slice/crio-84ed65e98f36228b023f4cec7d45da85d3f488338119c2f760b235de1fad8ae1 WatchSource:0}: Error finding container 84ed65e98f36228b023f4cec7d45da85d3f488338119c2f760b235de1fad8ae1: Status 404 returned error can't find the container with id 84ed65e98f36228b023f4cec7d45da85d3f488338119c2f760b235de1fad8ae1 Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.171767 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.174750 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38819d0b-58b8-429b-9969-5309efa8d187-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.192339 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.212275 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.217257 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-images\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.232141 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.243512 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-proxy-tls\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.251971 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.264123 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.264284 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.764256983 +0000 UTC m=+143.368595598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.264697 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.264994 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.764983102 +0000 UTC m=+143.369321697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.273634 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.299355 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.310109 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57632601-15a1-4aed-a66b-3f01331cb637-proxy-tls\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.330564 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnpbm\" (UniqueName: \"kubernetes.io/projected/8f527f0c-5a97-49b2-afe6-19e5fa9bbddb-kube-api-access-bnpbm\") pod \"openshift-config-operator-7777fb866f-pp9dw\" (UID: \"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.350145 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxxdb\" (UniqueName: \"kubernetes.io/projected/9246bb69-d48c-40bb-bb3f-1a02872bf250-kube-api-access-hxxdb\") pod \"openshift-apiserver-operator-796bbdcf4f-wd6k6\" (UID: \"9246bb69-d48c-40bb-bb3f-1a02872bf250\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.366015 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.366282 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.866249964 +0000 UTC m=+143.470588569 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.366553 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.366895 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.866882112 +0000 UTC m=+143.471220717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.370534 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adbace7e-3d6e-43d5-9331-af5f0b57fe20-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.389901 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twj7q\" (UniqueName: \"kubernetes.io/projected/048ff883-805e-4a2e-b939-86de1aa2e6e2-kube-api-access-twj7q\") pod \"downloads-7954f5f757-l69gd\" (UID: \"048ff883-805e-4a2e-b939-86de1aa2e6e2\") " pod="openshift-console/downloads-7954f5f757-l69gd" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.408221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvkkl\" (UniqueName: \"kubernetes.io/projected/a9b8f278-7174-4b90-b14f-6091cacdb5c2-kube-api-access-fvkkl\") pod \"apiserver-7bbb656c7d-rgp8x\" (UID: \"a9b8f278-7174-4b90-b14f-6091cacdb5c2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.412135 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.427424 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/764f8575-73dd-4450-aefe-0890671a8f98-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l8ft4\" (UID: \"764f8575-73dd-4450-aefe-0890671a8f98\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.445672 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qddmw\" (UniqueName: \"kubernetes.io/projected/d460a3fd-a303-4c45-85ba-23241d0e8976-kube-api-access-qddmw\") pod \"etcd-operator-b45778765-rv5w5\" (UID: \"d460a3fd-a303-4c45-85ba-23241d0e8976\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.468145 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.468266 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.968225735 +0000 UTC m=+143.572564350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.468514 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.468978 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:21.968970485 +0000 UTC m=+143.573309090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.469014 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-l69gd" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.469061 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9q7s\" (UniqueName: \"kubernetes.io/projected/adbace7e-3d6e-43d5-9331-af5f0b57fe20-kube-api-access-f9q7s\") pod \"cluster-image-registry-operator-dc59b4c8b-sb8m7\" (UID: \"adbace7e-3d6e-43d5-9331-af5f0b57fe20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.471307 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.475870 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.483481 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.490562 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.493873 4735 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.496431 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.504727 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.512143 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.531977 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.537686 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1296f167-e81f-4641-b1b0-d0de2b031d63-certs\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.551585 4735 request.go:700] Waited for 1.899357118s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.553791 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.570092 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.570654 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.070640038 +0000 UTC m=+143.674978643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.572766 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.580067 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1296f167-e81f-4641-b1b0-d0de2b031d63-node-bootstrap-token\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.620109 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.620368 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.629273 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2cf18025-0832-4850-8568-9af2a131711c-config-volume\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.632064 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.646235 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2cf18025-0832-4850-8568-9af2a131711c-metrics-tls\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.651789 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.673371 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.673912 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.173890904 +0000 UTC m=+143.778229579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.673986 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.692425 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.712099 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.716451 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x"] Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.732726 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e9f768a-e958-41b3-975e-45f565437df6-cert\") pod \"ingress-canary-xkl5g\" (UID: \"0e9f768a-e958-41b3-975e-45f565437df6\") " pod="openshift-ingress-canary/ingress-canary-xkl5g" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.772146 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktgps\" (UniqueName: \"kubernetes.io/projected/5b742605-ede0-46c6-add0-01cda9fa184d-kube-api-access-ktgps\") pod \"authentication-operator-69f744f599-8vp7b\" (UID: \"5b742605-ede0-46c6-add0-01cda9fa184d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.775874 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.775978 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.275956318 +0000 UTC m=+143.880294933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.776187 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.776708 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.276697738 +0000 UTC m=+143.881036343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.791836 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7hhj\" (UniqueName: \"kubernetes.io/projected/5ddc6014-cf08-45d5-8cc9-56fdfd119657-kube-api-access-d7hhj\") pod \"dns-operator-744455d44c-mtqwx\" (UID: \"5ddc6014-cf08-45d5-8cc9-56fdfd119657\") " pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.807240 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-l69gd"] Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.810001 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zlj6\" (UniqueName: \"kubernetes.io/projected/2d841f5e-d73a-4777-acc1-7b28f247e829-kube-api-access-4zlj6\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4r2r\" (UID: \"2d841f5e-d73a-4777-acc1-7b28f247e829\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.810674 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.821532 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.825110 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6"] Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.832218 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjh2j\" (UniqueName: \"kubernetes.io/projected/5dea56bb-e794-479f-8269-e31474c18920-kube-api-access-tjh2j\") pod \"console-operator-58897d9998-6c285\" (UID: \"5dea56bb-e794-479f-8269-e31474c18920\") " pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.849656 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h58m\" (UniqueName: \"kubernetes.io/projected/d417a7ee-7687-4086-8047-39e259a9bfbe-kube-api-access-6h58m\") pod \"console-f9d7485db-sjvm9\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.866672 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-bound-sa-token\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.877727 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.878178 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.378148734 +0000 UTC m=+143.982487339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.890264 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlfq7\" (UniqueName: \"kubernetes.io/projected/e095b65e-6f13-4a89-8418-78f5bda82cb5-kube-api-access-dlfq7\") pod \"controller-manager-879f6c89f-zqccv\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.905191 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf2hq\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-kube-api-access-cf2hq\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.926688 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4fvp\" (UniqueName: \"kubernetes.io/projected/8059f3ce-ca39-493d-a2a9-3ec71b2f7d15-kube-api-access-w4fvp\") pod \"catalog-operator-68c6474976-pnhjv\" (UID: \"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.956399 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/674120ce-7924-489f-ab5c-f39b39df718f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zkgjz\" (UID: \"674120ce-7924-489f-ab5c-f39b39df718f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.964659 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.978068 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03719f40-3ec6-416a-ba6e-33e7ebd963c6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pf2d7\" (UID: \"03719f40-3ec6-416a-ba6e-33e7ebd963c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.979111 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:21 crc kubenswrapper[4735]: E1122 08:05:21.979416 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.479385705 +0000 UTC m=+144.083724310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:21 crc kubenswrapper[4735]: I1122 08:05:21.984943 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c22gx\" (UniqueName: \"kubernetes.io/projected/38819d0b-58b8-429b-9969-5309efa8d187-kube-api-access-c22gx\") pod \"kube-storage-version-migrator-operator-b67b599dd-njhxg\" (UID: \"38819d0b-58b8-429b-9969-5309efa8d187\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.003922 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlg27\" (UniqueName: \"kubernetes.io/projected/9f713337-be7d-47e8-bc15-76b5795d3c43-kube-api-access-jlg27\") pod \"migrator-59844c95c7-dl4cp\" (UID: \"9f713337-be7d-47e8-bc15-76b5795d3c43\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.027356 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc9cf\" (UniqueName: \"kubernetes.io/projected/31a557ae-6002-49fd-a9af-b3ea23dfd952-kube-api-access-sc9cf\") pod \"collect-profiles-29396640-xdk4q\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.048639 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.052359 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7dn\" (UniqueName: \"kubernetes.io/projected/a1f75b9e-1a13-4f13-8a78-98111d8dfeaf-kube-api-access-wf7dn\") pod \"package-server-manager-789f6589d5-g5sst\" (UID: \"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.056022 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.060564 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rv5w5"] Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.062387 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.063452 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4"] Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.071061 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb5ct\" (UniqueName: \"kubernetes.io/projected/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-kube-api-access-fb5ct\") pod \"marketplace-operator-79b997595-v7k7n\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.072622 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw"] Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.077776 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-l69gd" event={"ID":"048ff883-805e-4a2e-b939-86de1aa2e6e2","Type":"ContainerStarted","Data":"a99996582ad4346eb44e471723eb3b940489a78e7fab72b04aeb591332a40d4d"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.078274 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7"] Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.079492 4735 generic.go:334] "Generic (PLEG): container finished" podID="994a0303-2bc7-4487-9b70-2733fe287a39" containerID="4f12a4458d8f4ee92eb8c9397a77b1a070dca27ec12f1360b90193688e72589d" exitCode=0 Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.079559 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" event={"ID":"994a0303-2bc7-4487-9b70-2733fe287a39","Type":"ContainerDied","Data":"4f12a4458d8f4ee92eb8c9397a77b1a070dca27ec12f1360b90193688e72589d"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.080295 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.080587 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.580568305 +0000 UTC m=+144.184906910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.080659 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.080680 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" event={"ID":"a9b8f278-7174-4b90-b14f-6091cacdb5c2","Type":"ContainerStarted","Data":"9a65892b02a99794299d749d46435daf8f5b3c9605f2ff77a7f5fac112461d43"} Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.080904 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.580896524 +0000 UTC m=+144.185235129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.082423 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" event={"ID":"510ea1b6-d5f7-42eb-b490-6f729076dcc3","Type":"ContainerStarted","Data":"3c85babd8d076ea19bdda177f82a697bef8ffd13979e2ba9cea31be198539422"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.082449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" event={"ID":"510ea1b6-d5f7-42eb-b490-6f729076dcc3","Type":"ContainerStarted","Data":"dc5aa9dc7124312a385abfe624ae051446511cae89104652abd71b30c05d25b7"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.082505 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" event={"ID":"510ea1b6-d5f7-42eb-b490-6f729076dcc3","Type":"ContainerStarted","Data":"84ed65e98f36228b023f4cec7d45da85d3f488338119c2f760b235de1fad8ae1"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.086194 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" event={"ID":"e64c4339-943c-48a4-9809-ad87c03cd233","Type":"ContainerStarted","Data":"12277d3f31e81f401dfa16753356edb84d9eab2805f3f74b5d50569b14c99a23"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.086225 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" event={"ID":"e64c4339-943c-48a4-9809-ad87c03cd233","Type":"ContainerStarted","Data":"83adf897901ea5865578249d194faf5ae7bddc78af84e92057395454fd8df668"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.086235 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" event={"ID":"e64c4339-943c-48a4-9809-ad87c03cd233","Type":"ContainerStarted","Data":"a401a59904fb4b4081f0ade207e6692e57655baec4094b65bf0b919f18aaa7cd"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.089115 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" event={"ID":"bcc3f64f-86dd-459c-8921-5cd3cf704ea5","Type":"ContainerStarted","Data":"8d3a30235bf3bd058ca8e4652aa583e12f2d92daf7b99a8392da299993bc6ac0"} Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.095738 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.096578 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a47af7c4-72b1-45bc-95c4-5f3207307fa2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.114510 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjdtq\" (UniqueName: \"kubernetes.io/projected/a24168d2-50e3-4ee4-bcbf-7abf4b57152e-kube-api-access-xjdtq\") pod \"olm-operator-6b444d44fb-9v46s\" (UID: \"a24168d2-50e3-4ee4-bcbf-7abf4b57152e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.126214 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.129666 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6l48\" (UniqueName: \"kubernetes.io/projected/2cf18025-0832-4850-8568-9af2a131711c-kube-api-access-q6l48\") pod \"dns-default-5s2w9\" (UID: \"2cf18025-0832-4850-8568-9af2a131711c\") " pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.146842 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjlcg\" (UniqueName: \"kubernetes.io/projected/53cd51d1-1576-492a-ba5e-e26a388b6a57-kube-api-access-cjlcg\") pod \"router-default-5444994796-w5v4q\" (UID: \"53cd51d1-1576-492a-ba5e-e26a388b6a57\") " pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.154974 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:22 crc kubenswrapper[4735]: W1122 08:05:22.156046 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9246bb69_d48c_40bb_bb3f_1a02872bf250.slice/crio-78a34356db4390c423a311b8fff4bf8443af4e6fc0cd34a80f47cd689dff5a4e WatchSource:0}: Error finding container 78a34356db4390c423a311b8fff4bf8443af4e6fc0cd34a80f47cd689dff5a4e: Status 404 returned error can't find the container with id 78a34356db4390c423a311b8fff4bf8443af4e6fc0cd34a80f47cd689dff5a4e Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.169555 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5r5w\" (UniqueName: \"kubernetes.io/projected/248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9-kube-api-access-s5r5w\") pod \"service-ca-operator-777779d784-xksr2\" (UID: \"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.169796 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.183507 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.183842 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.683818841 +0000 UTC m=+144.288157446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.183965 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.186026 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.686016841 +0000 UTC m=+144.290355446 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.190784 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdhrz\" (UniqueName: \"kubernetes.io/projected/a47af7c4-72b1-45bc-95c4-5f3207307fa2-kube-api-access-fdhrz\") pod \"ingress-operator-5b745b69d9-qg8pw\" (UID: \"a47af7c4-72b1-45bc-95c4-5f3207307fa2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.194928 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.207483 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.208090 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6hhs\" (UniqueName: \"kubernetes.io/projected/2da8d731-f76e-472f-bb05-b2242fb9a652-kube-api-access-n6hhs\") pod \"route-controller-manager-6576b87f9c-5rsm4\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.215859 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.224757 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.225152 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n5z9\" (UniqueName: \"kubernetes.io/projected/57632601-15a1-4aed-a66b-3f01331cb637-kube-api-access-5n5z9\") pod \"machine-config-controller-84d6567774-x8zj9\" (UID: \"57632601-15a1-4aed-a66b-3f01331cb637\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.243658 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.252528 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.253531 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfzhx\" (UniqueName: \"kubernetes.io/projected/bf91eba7-b4d7-486b-98e1-2ca53735adc5-kube-api-access-tfzhx\") pod \"service-ca-9c57cc56f-zm9cl\" (UID: \"bf91eba7-b4d7-486b-98e1-2ca53735adc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.259928 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.267389 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.270171 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpcmx\" (UniqueName: \"kubernetes.io/projected/0e9f768a-e958-41b3-975e-45f565437df6-kube-api-access-lpcmx\") pod \"ingress-canary-xkl5g\" (UID: \"0e9f768a-e958-41b3-975e-45f565437df6\") " pod="openshift-ingress-canary/ingress-canary-xkl5g" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.285124 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.285875 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.785853984 +0000 UTC m=+144.390192599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.288235 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqr46\" (UniqueName: \"kubernetes.io/projected/e8094c96-1edb-431d-8081-4a4927be57a8-kube-api-access-fqr46\") pod \"packageserver-d55dfcdfc-wmjbb\" (UID: \"e8094c96-1edb-431d-8081-4a4927be57a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.307347 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82nm9\" (UniqueName: \"kubernetes.io/projected/e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9-kube-api-access-82nm9\") pod \"machine-config-operator-74547568cd-5j6k5\" (UID: \"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.320864 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.325633 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xkl5g" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.332068 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl62v\" (UniqueName: \"kubernetes.io/projected/1296f167-e81f-4641-b1b0-d0de2b031d63-kube-api-access-hl62v\") pod \"machine-config-server-wrfpv\" (UID: \"1296f167-e81f-4641-b1b0-d0de2b031d63\") " pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.334773 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.339352 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8vp7b"] Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.346184 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7zlw\" (UniqueName: \"kubernetes.io/projected/4b91e784-c670-41af-95c6-696a7197e01a-kube-api-access-n7zlw\") pod \"multus-admission-controller-857f4d67dd-7w2f5\" (UID: \"4b91e784-c670-41af-95c6-696a7197e01a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.366525 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xhkl\" (UniqueName: \"kubernetes.io/projected/fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d-kube-api-access-5xhkl\") pod \"csi-hostpathplugin-4t74q\" (UID: \"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d\") " pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.386739 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.387086 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.887072734 +0000 UTC m=+144.491411339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.394114 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kwhz\" (UniqueName: \"kubernetes.io/projected/fb194fbb-af18-420c-ad19-004eba55e639-kube-api-access-7kwhz\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzwms\" (UID: \"fb194fbb-af18-420c-ad19-004eba55e639\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.395157 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r"] Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.440638 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.447356 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.461361 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.474519 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7"] Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.477359 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.488173 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.488244 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.488715 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:22.988685295 +0000 UTC m=+144.593023900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.525734 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zqccv"] Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.533279 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.578246 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.590692 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.591055 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.091037757 +0000 UTC m=+144.695376362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.602495 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4t74q" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.610198 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wrfpv" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.692510 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.692898 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.192883874 +0000 UTC m=+144.797222479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: W1122 08:05:22.717185 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53cd51d1_1576_492a_ba5e_e26a388b6a57.slice/crio-859fad9f7c27058517130f64403161493313ac4178a4895b239b3f0c9d8789bf WatchSource:0}: Error finding container 859fad9f7c27058517130f64403161493313ac4178a4895b239b3f0c9d8789bf: Status 404 returned error can't find the container with id 859fad9f7c27058517130f64403161493313ac4178a4895b239b3f0c9d8789bf Nov 22 08:05:22 crc kubenswrapper[4735]: W1122 08:05:22.731336 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode095b65e_6f13_4a89_8418_78f5bda82cb5.slice/crio-3075a3c5776922c438d94b69f2819c1cf146fd543a7148c67342576a310e4eed WatchSource:0}: Error finding container 3075a3c5776922c438d94b69f2819c1cf146fd543a7148c67342576a310e4eed: Status 404 returned error can't find the container with id 3075a3c5776922c438d94b69f2819c1cf146fd543a7148c67342576a310e4eed Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.747085 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.793740 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.794262 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.294243738 +0000 UTC m=+144.898582333 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.895107 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.895275 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.395240233 +0000 UTC m=+144.999578858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.895340 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.895662 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.395648124 +0000 UTC m=+144.999986729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.997164 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.997358 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.497331377 +0000 UTC m=+145.101669982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:22 crc kubenswrapper[4735]: I1122 08:05:22.997461 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:22 crc kubenswrapper[4735]: E1122 08:05:22.997752 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.497736708 +0000 UTC m=+145.102075313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.099773 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-l69gd" event={"ID":"048ff883-805e-4a2e-b939-86de1aa2e6e2","Type":"ContainerStarted","Data":"dcb63e79ef37ca9ec94ad944f19ae3c0b77617f8eb865f3f68bf403efe7e13ea"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.100103 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.100403 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.600383588 +0000 UTC m=+145.204722193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.116083 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" event={"ID":"e095b65e-6f13-4a89-8418-78f5bda82cb5","Type":"ContainerStarted","Data":"3075a3c5776922c438d94b69f2819c1cf146fd543a7148c67342576a310e4eed"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.148396 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-98dz8" podStartSLOduration=123.148375737 podStartE2EDuration="2m3.148375737s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:23.147189624 +0000 UTC m=+144.751528239" watchObservedRunningTime="2025-11-22 08:05:23.148375737 +0000 UTC m=+144.752714342" Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.170729 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" event={"ID":"9246bb69-d48c-40bb-bb3f-1a02872bf250","Type":"ContainerStarted","Data":"d031b5297b9ae17e5c7215336250a1a32da64b158ef4f66dac430148f60d7381"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.170774 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" event={"ID":"9246bb69-d48c-40bb-bb3f-1a02872bf250","Type":"ContainerStarted","Data":"78a34356db4390c423a311b8fff4bf8443af4e6fc0cd34a80f47cd689dff5a4e"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.176330 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wrfpv" event={"ID":"1296f167-e81f-4641-b1b0-d0de2b031d63","Type":"ContainerStarted","Data":"f53c4c45a676dad8f70294655319eaac0c97d59a1bb8b1f56a6723d7a221d4fc"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.200298 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" event={"ID":"5b742605-ede0-46c6-add0-01cda9fa184d","Type":"ContainerStarted","Data":"727ec201759a7a8ed6bc59d10924e96acb233ecffd9e29cac8b0fda41c00d4b8"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.201147 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.201471 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.701460304 +0000 UTC m=+145.305798909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.203745 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" event={"ID":"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb","Type":"ContainerStarted","Data":"4690b02df29f91deb3a7e7cda9435723c9d9cd53c2c9ad3b8cdfdc3cb3200f93"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.210310 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s"] Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.210938 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mtqwx"] Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.214955 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" event={"ID":"d460a3fd-a303-4c45-85ba-23241d0e8976","Type":"ContainerStarted","Data":"40aa3b998fe74524bfc841ea8588bd7e7aabadddc86a876052875d30d5d74664"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.218573 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" event={"ID":"03719f40-3ec6-416a-ba6e-33e7ebd963c6","Type":"ContainerStarted","Data":"ec23a437371176a891709f525ec6cab8571c7a37b078d49eb63ace10aea48e20"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.221199 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" event={"ID":"2d841f5e-d73a-4777-acc1-7b28f247e829","Type":"ContainerStarted","Data":"d74ccb4dcf8aed8de94957dbab9897cb5206141cbf9f8443a70b553f586057bc"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.223867 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w5v4q" event={"ID":"53cd51d1-1576-492a-ba5e-e26a388b6a57","Type":"ContainerStarted","Data":"859fad9f7c27058517130f64403161493313ac4178a4895b239b3f0c9d8789bf"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.225095 4735 generic.go:334] "Generic (PLEG): container finished" podID="a9b8f278-7174-4b90-b14f-6091cacdb5c2" containerID="39066c6e517935d384d60f553a1f4d66b40ab9d205872bd63cc61282427cc513" exitCode=0 Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.225150 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" event={"ID":"a9b8f278-7174-4b90-b14f-6091cacdb5c2","Type":"ContainerDied","Data":"39066c6e517935d384d60f553a1f4d66b40ab9d205872bd63cc61282427cc513"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.225711 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" event={"ID":"764f8575-73dd-4450-aefe-0890671a8f98","Type":"ContainerStarted","Data":"aacc4180847b81c21af1a04c4d710991a0c081d00da62556ba15ba2b92a61334"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.227200 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" event={"ID":"adbace7e-3d6e-43d5-9331-af5f0b57fe20","Type":"ContainerStarted","Data":"b625b5993e29132165d78496cae40b77c8ed07a47fc2be6d6937f8ee0dffe377"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.227246 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" event={"ID":"adbace7e-3d6e-43d5-9331-af5f0b57fe20","Type":"ContainerStarted","Data":"6f05be532b0fb3e0b63884ec7016535ffe4fb613d316b7ef9965c5177a2d82e3"} Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.303219 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.304714 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.804688919 +0000 UTC m=+145.409027524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.308188 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.309824 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.809812869 +0000 UTC m=+145.414151474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.408940 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.409344 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:23.909307392 +0000 UTC m=+145.513645997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.461020 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv"] Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.475341 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sjvm9"] Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.511578 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.511944 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.011927451 +0000 UTC m=+145.616266066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.615098 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.615480 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.115465664 +0000 UTC m=+145.719804269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: W1122 08:05:23.651796 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd417a7ee_7687_4086_8047_39e259a9bfbe.slice/crio-ef3b3445da7cdc478758902e9bb13787348f4acb5cf03b8afcdc356c391b8b8e WatchSource:0}: Error finding container ef3b3445da7cdc478758902e9bb13787348f4acb5cf03b8afcdc356c391b8b8e: Status 404 returned error can't find the container with id ef3b3445da7cdc478758902e9bb13787348f4acb5cf03b8afcdc356c391b8b8e Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.716808 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.717601 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.217548239 +0000 UTC m=+145.821886844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.820143 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.835882 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.335857365 +0000 UTC m=+145.940195970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.839065 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.839668 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.339639658 +0000 UTC m=+145.943978263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.914416 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wdh9p" podStartSLOduration=123.914399337 podStartE2EDuration="2m3.914399337s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:23.913782821 +0000 UTC m=+145.518121426" watchObservedRunningTime="2025-11-22 08:05:23.914399337 +0000 UTC m=+145.518737962" Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.947240 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:23 crc kubenswrapper[4735]: E1122 08:05:23.947686 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.447667335 +0000 UTC m=+146.052005940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:23 crc kubenswrapper[4735]: I1122 08:05:23.962664 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w972q" podStartSLOduration=123.962640933 podStartE2EDuration="2m3.962640933s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:23.95812965 +0000 UTC m=+145.562468255" watchObservedRunningTime="2025-11-22 08:05:23.962640933 +0000 UTC m=+145.566979538" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.049223 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.049581 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.549569103 +0000 UTC m=+146.153907708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.108538 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp"] Nov 22 08:05:24 crc kubenswrapper[4735]: W1122 08:05:24.128980 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f713337_be7d_47e8_bc15_76b5795d3c43.slice/crio-bf02ac0e805b682f320a7f7f9f3e03dfdcde84187cf123c31a9578051515674f WatchSource:0}: Error finding container bf02ac0e805b682f320a7f7f9f3e03dfdcde84187cf123c31a9578051515674f: Status 404 returned error can't find the container with id bf02ac0e805b682f320a7f7f9f3e03dfdcde84187cf123c31a9578051515674f Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.130030 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" podStartSLOduration=124.130009007 podStartE2EDuration="2m4.130009007s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.117044393 +0000 UTC m=+145.721382998" watchObservedRunningTime="2025-11-22 08:05:24.130009007 +0000 UTC m=+145.734347612" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.149855 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6c285"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.150737 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.151083 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.651067452 +0000 UTC m=+146.255406057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.159304 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q"] Nov 22 08:05:24 crc kubenswrapper[4735]: W1122 08:05:24.164167 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31a557ae_6002_49fd_a9af_b3ea23dfd952.slice/crio-a3577f7bfba6b019d9c0ee4100f9e668ca6567238bd3ded96c2a5136123a2d01 WatchSource:0}: Error finding container a3577f7bfba6b019d9c0ee4100f9e668ca6567238bd3ded96c2a5136123a2d01: Status 404 returned error can't find the container with id a3577f7bfba6b019d9c0ee4100f9e668ca6567238bd3ded96c2a5136123a2d01 Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.164221 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.198005 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7k7n"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.244655 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.244721 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" event={"ID":"38819d0b-58b8-429b-9969-5309efa8d187","Type":"ContainerStarted","Data":"4e82c4489570e98471c7441f1615fa6057c1f29fc353b651d77cd391925cb2d1"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.247169 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" event={"ID":"a24168d2-50e3-4ee4-bcbf-7abf4b57152e","Type":"ContainerStarted","Data":"a940807de93afb693f5629b8ba483b3cfd15c9785a23bfb810d1ec00f943dab1"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.247285 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" event={"ID":"a24168d2-50e3-4ee4-bcbf-7abf4b57152e","Type":"ContainerStarted","Data":"605296e85ca8c18c6dc1e4c5c9c8c8efb664333c0280f6f95dd44ec2e7fec266"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.248014 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sb8m7" podStartSLOduration=124.247989665 podStartE2EDuration="2m4.247989665s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.209078613 +0000 UTC m=+145.813417218" watchObservedRunningTime="2025-11-22 08:05:24.247989665 +0000 UTC m=+145.852328270" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.248287 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.254571 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.254956 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.754943505 +0000 UTC m=+146.359282100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.255326 4735 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9v46s container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.255373 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" podUID="a24168d2-50e3-4ee4-bcbf-7abf4b57152e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.263574 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.266683 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb"] Nov 22 08:05:24 crc kubenswrapper[4735]: W1122 08:05:24.268763 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1f75b9e_1a13_4f13_8a78_98111d8dfeaf.slice/crio-b00e23acd4c60b663435663579348188df8116d4abec78cbcdd1b86d583c87aa WatchSource:0}: Error finding container b00e23acd4c60b663435663579348188df8116d4abec78cbcdd1b86d583c87aa: Status 404 returned error can't find the container with id b00e23acd4c60b663435663579348188df8116d4abec78cbcdd1b86d583c87aa Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.269025 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.279173 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.279258 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xkl5g"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.279807 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" event={"ID":"5ddc6014-cf08-45d5-8cc9-56fdfd119657","Type":"ContainerStarted","Data":"1f39f2b0a6d437d35a5e67f442023238a30466082237c4a50e58cad9e907acfc"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.280040 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" event={"ID":"5ddc6014-cf08-45d5-8cc9-56fdfd119657","Type":"ContainerStarted","Data":"f7d5b07082576549cdc6c077e0c6c8b5447e6a2dcf32d301c9819e80167812fe"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.280053 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" event={"ID":"5ddc6014-cf08-45d5-8cc9-56fdfd119657","Type":"ContainerStarted","Data":"d855b14b73a5b02e49203278c55bcf71fa12028220f1846e4369fb80e756eae4"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.301906 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wd6k6" podStartSLOduration=124.301884414 podStartE2EDuration="2m4.301884414s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.280343898 +0000 UTC m=+145.884682503" watchObservedRunningTime="2025-11-22 08:05:24.301884414 +0000 UTC m=+145.906223019" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.312561 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.314140 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" event={"ID":"2d841f5e-d73a-4777-acc1-7b28f247e829","Type":"ContainerStarted","Data":"65d4ee891bc89a6b22d7bb416fe45e3dce74b6f723ffe66b23e20c0f14c5e6b7"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.317646 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6c285" event={"ID":"5dea56bb-e794-479f-8269-e31474c18920","Type":"ContainerStarted","Data":"2605d4ed06ba27927294775eca2bcfec2a8dfe553be6830f04cae6f795fac737"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.329049 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sjvm9" event={"ID":"d417a7ee-7687-4086-8047-39e259a9bfbe","Type":"ContainerStarted","Data":"fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.329112 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sjvm9" event={"ID":"d417a7ee-7687-4086-8047-39e259a9bfbe","Type":"ContainerStarted","Data":"ef3b3445da7cdc478758902e9bb13787348f4acb5cf03b8afcdc356c391b8b8e"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.329572 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xksr2"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.336762 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5s2w9"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.337930 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" event={"ID":"764f8575-73dd-4450-aefe-0890671a8f98","Type":"ContainerStarted","Data":"4f61b3f9395b4a748001993fe806a3c4a81f51cf0e31c3f439f8f9eb6157932a"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.362027 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.362598 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4t74q"] Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.378052 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.878017841 +0000 UTC m=+146.482356446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.401778 4735 generic.go:334] "Generic (PLEG): container finished" podID="8f527f0c-5a97-49b2-afe6-19e5fa9bbddb" containerID="a54623eed7cb22d008ecfa20379527f02df3e5e7774ef0880906e25a1e63ba4a" exitCode=0 Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.401892 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" event={"ID":"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb","Type":"ContainerDied","Data":"a54623eed7cb22d008ecfa20379527f02df3e5e7774ef0880906e25a1e63ba4a"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.441468 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" event={"ID":"994a0303-2bc7-4487-9b70-2733fe287a39","Type":"ContainerStarted","Data":"c69844dc7e6651b5b67e1af938dabc5648ccbadbb1719f8a373dfcfe60596b37"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.441552 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" event={"ID":"994a0303-2bc7-4487-9b70-2733fe287a39","Type":"ContainerStarted","Data":"53be6c10837f14ad0e7387292f4bd3f7893d86712e498bf20002a46cde211f80"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.466177 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-mtqwx" podStartSLOduration=124.466153884 podStartE2EDuration="2m4.466153884s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.424460967 +0000 UTC m=+146.028799572" watchObservedRunningTime="2025-11-22 08:05:24.466153884 +0000 UTC m=+146.070492489" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.466706 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.478283 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" podStartSLOduration=123.477089883 podStartE2EDuration="2m3.477089883s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.475055757 +0000 UTC m=+146.079394362" watchObservedRunningTime="2025-11-22 08:05:24.477089883 +0000 UTC m=+146.081428508" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.480607 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.481824 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:24.981804502 +0000 UTC m=+146.586143217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.485186 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7w2f5"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.490366 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w5v4q" event={"ID":"53cd51d1-1576-492a-ba5e-e26a388b6a57","Type":"ContainerStarted","Data":"cc46444aec4ebce85c6b0ab3b23f4eb48506784d172311d8f1e59ab3b617429a"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.497921 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" event={"ID":"9f713337-be7d-47e8-bc15-76b5795d3c43","Type":"ContainerStarted","Data":"bf02ac0e805b682f320a7f7f9f3e03dfdcde84187cf123c31a9578051515674f"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.501086 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" event={"ID":"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15","Type":"ContainerStarted","Data":"ab8e6a97fcec425a23fde740bacdf71e2669676f59e4c8f843b63cda146c034b"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.501146 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" event={"ID":"8059f3ce-ca39-493d-a2a9-3ec71b2f7d15","Type":"ContainerStarted","Data":"8a584e6fc8d9ab086d907da93105635593533f1e555dca23943826aa1c6c5b0a"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.502039 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.503080 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" event={"ID":"31a557ae-6002-49fd-a9af-b3ea23dfd952","Type":"ContainerStarted","Data":"a3577f7bfba6b019d9c0ee4100f9e668ca6567238bd3ded96c2a5136123a2d01"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.506356 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" event={"ID":"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3","Type":"ContainerStarted","Data":"515965c2af5d1e9d156c0ae801d8fd7d2a600fe4070d12da588fa722595094d0"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.512078 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" event={"ID":"03719f40-3ec6-416a-ba6e-33e7ebd963c6","Type":"ContainerStarted","Data":"240ae605736c83b80fa42dd1d16164b562df77c86e6de267c37f7ac6cae8396d"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.520210 4735 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pnhjv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.520429 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" podUID="8059f3ce-ca39-493d-a2a9-3ec71b2f7d15" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.534244 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" event={"ID":"5b742605-ede0-46c6-add0-01cda9fa184d","Type":"ContainerStarted","Data":"70aea2889785973474cc186d8fe47edc32b747e5a48000b0d943c446a324fe93"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.535443 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4r2r" podStartSLOduration=124.535422164 podStartE2EDuration="2m4.535422164s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.534086967 +0000 UTC m=+146.138425572" watchObservedRunningTime="2025-11-22 08:05:24.535422164 +0000 UTC m=+146.139760769" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.554521 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.555030 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" event={"ID":"e095b65e-6f13-4a89-8418-78f5bda82cb5","Type":"ContainerStarted","Data":"fccf5d6f15567cd849c87eedb7e31d04969c435541bbc3011e279c73ed013406"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.555526 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zm9cl"] Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.555780 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:24 crc kubenswrapper[4735]: W1122 08:05:24.559671 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod674120ce_7924_489f_ab5c_f39b39df718f.slice/crio-14de656937018c6b8bc586f21e612def7b995d30a4ed63ee50d05ba25f660a0f WatchSource:0}: Error finding container 14de656937018c6b8bc586f21e612def7b995d30a4ed63ee50d05ba25f660a0f: Status 404 returned error can't find the container with id 14de656937018c6b8bc586f21e612def7b995d30a4ed63ee50d05ba25f660a0f Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.559689 4735 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-zqccv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.559753 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" podUID="e095b65e-6f13-4a89-8418-78f5bda82cb5" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.582270 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.583064 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.083025362 +0000 UTC m=+146.687363967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.583390 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" event={"ID":"a9b8f278-7174-4b90-b14f-6091cacdb5c2","Type":"ContainerStarted","Data":"dda4b37d5130b9cedc9a081c5e156d871aa22d1d3d555259da7308220f15b33d"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.606287 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wrfpv" event={"ID":"1296f167-e81f-4641-b1b0-d0de2b031d63","Type":"ContainerStarted","Data":"3e98d1ad22f27dce614f8f324b8f4ba4418545337ef59ba484abdda82ef761b0"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.613280 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-w5v4q" podStartSLOduration=124.613261077 podStartE2EDuration="2m4.613261077s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.608878997 +0000 UTC m=+146.213217602" watchObservedRunningTime="2025-11-22 08:05:24.613261077 +0000 UTC m=+146.217599682" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.613612 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pf2d7" podStartSLOduration=124.613606666 podStartE2EDuration="2m4.613606666s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.565596387 +0000 UTC m=+146.169935002" watchObservedRunningTime="2025-11-22 08:05:24.613606666 +0000 UTC m=+146.217945281" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.628984 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" event={"ID":"d460a3fd-a303-4c45-85ba-23241d0e8976","Type":"ContainerStarted","Data":"22e60825cab840b073444ed117ec57f85a1ca21f342804edb60abbef8ee13a18"} Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.629043 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-l69gd" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.642782 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-sjvm9" podStartSLOduration=124.642762891 podStartE2EDuration="2m4.642762891s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.636157141 +0000 UTC m=+146.240495766" watchObservedRunningTime="2025-11-22 08:05:24.642762891 +0000 UTC m=+146.247101496" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.646736 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-l69gd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.646799 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-l69gd" podUID="048ff883-805e-4a2e-b939-86de1aa2e6e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.688633 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.692013 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.191993804 +0000 UTC m=+146.796332619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.717783 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l8ft4" podStartSLOduration=124.717766527 podStartE2EDuration="2m4.717766527s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.71713049 +0000 UTC m=+146.321469095" watchObservedRunningTime="2025-11-22 08:05:24.717766527 +0000 UTC m=+146.322105132" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.790844 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.791008 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.290981183 +0000 UTC m=+146.895319778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.791951 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" podStartSLOduration=123.791932909 podStartE2EDuration="2m3.791932909s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.744071274 +0000 UTC m=+146.348409879" watchObservedRunningTime="2025-11-22 08:05:24.791932909 +0000 UTC m=+146.396271514" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.793526 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" podStartSLOduration=124.793518953 podStartE2EDuration="2m4.793518953s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.786833491 +0000 UTC m=+146.391172106" watchObservedRunningTime="2025-11-22 08:05:24.793518953 +0000 UTC m=+146.397857558" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.793720 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.795107 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.295090925 +0000 UTC m=+146.899429530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.829755 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-wrfpv" podStartSLOduration=5.82973886 podStartE2EDuration="5.82973886s" podCreationTimestamp="2025-11-22 08:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.829451133 +0000 UTC m=+146.433789738" watchObservedRunningTime="2025-11-22 08:05:24.82973886 +0000 UTC m=+146.434077465" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.831575 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" podStartSLOduration=124.83156591 podStartE2EDuration="2m4.83156591s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.816364555 +0000 UTC m=+146.420703160" watchObservedRunningTime="2025-11-22 08:05:24.83156591 +0000 UTC m=+146.435904515" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.878029 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rv5w5" podStartSLOduration=124.878009897 podStartE2EDuration="2m4.878009897s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.877430321 +0000 UTC m=+146.481768926" watchObservedRunningTime="2025-11-22 08:05:24.878009897 +0000 UTC m=+146.482348492" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.887125 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-l69gd" podStartSLOduration=124.887108044 podStartE2EDuration="2m4.887108044s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.849008116 +0000 UTC m=+146.453346721" watchObservedRunningTime="2025-11-22 08:05:24.887108044 +0000 UTC m=+146.491446650" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.895019 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.895353 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.395321639 +0000 UTC m=+146.999660244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.895394 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:24 crc kubenswrapper[4735]: E1122 08:05:24.896351 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.396337067 +0000 UTC m=+147.000675672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.911793 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" podStartSLOduration=124.911764377 podStartE2EDuration="2m4.911764377s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.907287546 +0000 UTC m=+146.511626171" watchObservedRunningTime="2025-11-22 08:05:24.911764377 +0000 UTC m=+146.516102982" Nov 22 08:05:24 crc kubenswrapper[4735]: I1122 08:05:24.948742 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" podStartSLOduration=123.948720225 podStartE2EDuration="2m3.948720225s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:24.947178943 +0000 UTC m=+146.551517548" watchObservedRunningTime="2025-11-22 08:05:24.948720225 +0000 UTC m=+146.553058830" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.011243 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.011407 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.511363274 +0000 UTC m=+147.115701879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.012232 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.012908 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.512895815 +0000 UTC m=+147.117234440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.115650 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.116060 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.616016468 +0000 UTC m=+147.220355083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.220187 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.220673 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.720659111 +0000 UTC m=+147.324997716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.262716 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.264832 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:25 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:25 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:25 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.264865 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.321228 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.321958 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.821927754 +0000 UTC m=+147.426266359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.322021 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.322333 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.822326915 +0000 UTC m=+147.426665520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.423923 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.424658 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.924609634 +0000 UTC m=+147.528948239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.424755 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.425198 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:25.925177379 +0000 UTC m=+147.529516194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.546574 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.546908 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.046879819 +0000 UTC m=+147.651218434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.547157 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.547482 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.047470654 +0000 UTC m=+147.651809259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.647733 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.648126 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.148109149 +0000 UTC m=+147.752447754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.651449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" event={"ID":"2da8d731-f76e-472f-bb05-b2242fb9a652","Type":"ContainerStarted","Data":"e377a3308baafe60f81aa8f63d7e6d2c6553997e36c10002f44b98cb04b05cb0"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.651552 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" event={"ID":"2da8d731-f76e-472f-bb05-b2242fb9a652","Type":"ContainerStarted","Data":"29bfdbe49315178c37aed49f42ceab871fffb7a7d0b660f0317e5478b8f84a02"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.653243 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.653824 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.653875 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.658655 4735 patch_prober.go:28] interesting pod/apiserver-76f77b778f-rxw97 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.658729 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" podUID="994a0303-2bc7-4487-9b70-2733fe287a39" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.658655 4735 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5rsm4 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.658810 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" podUID="2da8d731-f76e-472f-bb05-b2242fb9a652" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.662900 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" event={"ID":"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3","Type":"ContainerStarted","Data":"20d3b007d12f2d1240dfdc25605e42412fe3b015a6507a07d52bf7ce9317d5f1"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.663086 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.666661 4735 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-v7k7n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.666724 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" podUID="beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.671132 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" event={"ID":"674120ce-7924-489f-ab5c-f39b39df718f","Type":"ContainerStarted","Data":"85eed55c21c59e359d7abc87ca8211565b85079427438a14d79f7747e9bbe935"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.671178 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" event={"ID":"674120ce-7924-489f-ab5c-f39b39df718f","Type":"ContainerStarted","Data":"14de656937018c6b8bc586f21e612def7b995d30a4ed63ee50d05ba25f660a0f"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.672014 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" podStartSLOduration=124.67200098 podStartE2EDuration="2m4.67200098s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:25.671392334 +0000 UTC m=+147.275730939" watchObservedRunningTime="2025-11-22 08:05:25.67200098 +0000 UTC m=+147.276339585" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.676351 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" event={"ID":"9f713337-be7d-47e8-bc15-76b5795d3c43","Type":"ContainerStarted","Data":"fb376c619c2d0499cd156e7bdd8ed49cd3e6e8c41e107f0eaa4e075f3ba43250"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.676916 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" event={"ID":"9f713337-be7d-47e8-bc15-76b5795d3c43","Type":"ContainerStarted","Data":"51c89faaef1952498a3b398b40240fad656872ca1a6fd6da4270756f6929e091"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.683941 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" event={"ID":"8f527f0c-5a97-49b2-afe6-19e5fa9bbddb","Type":"ContainerStarted","Data":"ee737ba68af655da4de12cff4381b433f95786f8ff9b1f22512eab8378d44593"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.684719 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.692561 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5s2w9" event={"ID":"2cf18025-0832-4850-8568-9af2a131711c","Type":"ContainerStarted","Data":"c36a771e3dfccfe12831eecefcfc4a11b2c591a2542a1d0aba1e1dd17f0b5338"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.692607 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5s2w9" event={"ID":"2cf18025-0832-4850-8568-9af2a131711c","Type":"ContainerStarted","Data":"a0c2c33b40a565d95ebff407b05644cfb9563ac58ed80ad0c72d91509b9a32a6"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.698146 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" podStartSLOduration=124.698129493 podStartE2EDuration="2m4.698129493s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:25.696660533 +0000 UTC m=+147.300999138" watchObservedRunningTime="2025-11-22 08:05:25.698129493 +0000 UTC m=+147.302468098" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.707331 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" event={"ID":"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9","Type":"ContainerStarted","Data":"efb34d0970cc44285b9c177fc94a3aab831983a3ffa836c569847fe4421cd1db"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.707376 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" event={"ID":"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9","Type":"ContainerStarted","Data":"da2fd069067ce8c217bb936fb1b287823b470c72cc77985a20fca972f82208e5"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.738021 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zkgjz" podStartSLOduration=125.738004131 podStartE2EDuration="2m5.738004131s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:25.736578892 +0000 UTC m=+147.340917507" watchObservedRunningTime="2025-11-22 08:05:25.738004131 +0000 UTC m=+147.342342736" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.745185 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" event={"ID":"e8094c96-1edb-431d-8081-4a4927be57a8","Type":"ContainerStarted","Data":"5c44a001d4669529405ba97df4b0d788975c80ca027ea4259f832137d94f217c"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.745244 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" event={"ID":"e8094c96-1edb-431d-8081-4a4927be57a8","Type":"ContainerStarted","Data":"95f32d12194a30508ff80267a6faacb99dc76140582f2b3038241091710e65ce"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.746437 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.748555 4735 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wmjbb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.748593 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" podUID="e8094c96-1edb-431d-8081-4a4927be57a8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.749613 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.753686 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.253668208 +0000 UTC m=+147.858006893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.757496 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t74q" event={"ID":"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d","Type":"ContainerStarted","Data":"b48191e77e3eab2dcd9b0f1979f873b97158c678dcf3c8881a719fbebfc47fb3"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.759953 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" event={"ID":"fb194fbb-af18-420c-ad19-004eba55e639","Type":"ContainerStarted","Data":"4df3ea32ab9a5ec42f4e22ee26de31fdfe116f0b8c2bcedfaac01e4d26b17903"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.759984 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" event={"ID":"fb194fbb-af18-420c-ad19-004eba55e639","Type":"ContainerStarted","Data":"beb0e5b6bfdf0734e05238890916539c8de6be9853df9bc3d3014fe3ffb1f3e8"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.761842 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" event={"ID":"a47af7c4-72b1-45bc-95c4-5f3207307fa2","Type":"ContainerStarted","Data":"272e8c518a3d1be30411c03713cf0e9649b4fcce8b3202beb17619703169d1a6"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.761866 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" event={"ID":"a47af7c4-72b1-45bc-95c4-5f3207307fa2","Type":"ContainerStarted","Data":"7a09e505130c89692f12346f1142cc0bbcc8caf800b0d7a16cf451d49bcc35fc"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.778355 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" event={"ID":"4b91e784-c670-41af-95c6-696a7197e01a","Type":"ContainerStarted","Data":"43955cd5c2bbba1b647595ec59a66ab39696d383face4258783ae0d4f558e2f7"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.778400 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" event={"ID":"4b91e784-c670-41af-95c6-696a7197e01a","Type":"ContainerStarted","Data":"2316abaab9f2f61d671e552b3d0b3a2b9e2c8b8654589520e38727b2a56da037"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.789748 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" event={"ID":"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf","Type":"ContainerStarted","Data":"18928d3e2da32afd23d2d79e1f1870ba048d12b89cee86a2d843760b8714c3c8"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.789792 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" event={"ID":"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf","Type":"ContainerStarted","Data":"4cc7f58435a9cd43217aace70dea0ab068302b8b926a411044b9872af063ed05"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.789824 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" event={"ID":"a1f75b9e-1a13-4f13-8a78-98111d8dfeaf","Type":"ContainerStarted","Data":"b00e23acd4c60b663435663579348188df8116d4abec78cbcdd1b86d583c87aa"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.790313 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.803454 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6c285" event={"ID":"5dea56bb-e794-479f-8269-e31474c18920","Type":"ContainerStarted","Data":"115625169be578d98c4c582ac5e173cadbe3c0d40562b6ba321c35152add5477"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.805186 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.810706 4735 patch_prober.go:28] interesting pod/console-operator-58897d9998-6c285 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.810758 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-6c285" podUID="5dea56bb-e794-479f-8269-e31474c18920" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.829774 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" event={"ID":"bf91eba7-b4d7-486b-98e1-2ca53735adc5","Type":"ContainerStarted","Data":"445420bc4f6e74d782762e72eb8a2e2e0e65cf09936019fa74be4d049143b759"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.829836 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" event={"ID":"bf91eba7-b4d7-486b-98e1-2ca53735adc5","Type":"ContainerStarted","Data":"476056ac6057ca2950ee96b55eaafe7610233b2e6d409f50d39b4a6f8adfc437"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.833163 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dl4cp" podStartSLOduration=125.833136955 podStartE2EDuration="2m5.833136955s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:25.830689038 +0000 UTC m=+147.435027643" watchObservedRunningTime="2025-11-22 08:05:25.833136955 +0000 UTC m=+147.437475560" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.842393 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" podStartSLOduration=125.842357426 podStartE2EDuration="2m5.842357426s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:25.790200915 +0000 UTC m=+147.394539520" watchObservedRunningTime="2025-11-22 08:05:25.842357426 +0000 UTC m=+147.446696031" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.851987 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.852661 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.352644657 +0000 UTC m=+147.956983262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.857559 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" event={"ID":"57632601-15a1-4aed-a66b-3f01331cb637","Type":"ContainerStarted","Data":"90a621d3e66e6fef054711655f9b7ba8b8c61b314c7a035f3138ad418e9d288c"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.857616 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" event={"ID":"57632601-15a1-4aed-a66b-3f01331cb637","Type":"ContainerStarted","Data":"6b8b042238f90aebfc8cfdd262ab4559765233a81c1f1b24f74fec2a8323c184"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.857642 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" event={"ID":"57632601-15a1-4aed-a66b-3f01331cb637","Type":"ContainerStarted","Data":"451fa857a451ffb62d60268336d20ba1b12d4b9ff04c280200eef3ae9ef2c5a4"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.883349 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" podStartSLOduration=125.883318964 podStartE2EDuration="2m5.883318964s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:25.870911085 +0000 UTC m=+147.475249690" watchObservedRunningTime="2025-11-22 08:05:25.883318964 +0000 UTC m=+147.487657569" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.889284 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" event={"ID":"38819d0b-58b8-429b-9969-5309efa8d187","Type":"ContainerStarted","Data":"cfeb3d99dad69a8ee4d3a6debcab88e5b9e45e3ce01f3048198b09bb1c4aafaf"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.913792 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xkl5g" event={"ID":"0e9f768a-e958-41b3-975e-45f565437df6","Type":"ContainerStarted","Data":"2f111dacab06a9df2f6ff4c197a7096b28ef2468733b17082115ae61e2b3e779"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.913860 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xkl5g" event={"ID":"0e9f768a-e958-41b3-975e-45f565437df6","Type":"ContainerStarted","Data":"2f530f4225a0d34cd49e784c45d3ff243ccab0932a03cbde872c3b65363577c9"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.916790 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" event={"ID":"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9","Type":"ContainerStarted","Data":"4fb9977e742d897965683bd45392adae1caae6b08321e3613b863671c8680692"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.916855 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" event={"ID":"248ca8a7-eed5-45ee-bb4a-c917bc8cb5c9","Type":"ContainerStarted","Data":"bec8644494a25435cf09bd708ea5073c966c4252b68058e43fb36e9f16d5eb29"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.920106 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" event={"ID":"31a557ae-6002-49fd-a9af-b3ea23dfd952","Type":"ContainerStarted","Data":"edf114b0a6ecb80547997ca80f1020152ca9adb4047e825a975d92c607a1f28c"} Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.924814 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-l69gd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.924860 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-l69gd" podUID="048ff883-805e-4a2e-b939-86de1aa2e6e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.938799 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.977857 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" podStartSLOduration=124.977831592 podStartE2EDuration="2m4.977831592s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:25.904949864 +0000 UTC m=+147.509288479" watchObservedRunningTime="2025-11-22 08:05:25.977831592 +0000 UTC m=+147.582170197" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.978505 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-6c285" podStartSLOduration=125.97850115 podStartE2EDuration="2m5.97850115s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:25.940134223 +0000 UTC m=+147.544472828" watchObservedRunningTime="2025-11-22 08:05:25.97850115 +0000 UTC m=+147.582839755" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.984525 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pnhjv" Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.986029 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:25 crc kubenswrapper[4735]: E1122 08:05:25.991810 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.491787802 +0000 UTC m=+148.096126407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:25 crc kubenswrapper[4735]: I1122 08:05:25.997464 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.073762 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zm9cl" podStartSLOduration=125.073741037 podStartE2EDuration="2m5.073741037s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.071891296 +0000 UTC m=+147.676229901" watchObservedRunningTime="2025-11-22 08:05:26.073741037 +0000 UTC m=+147.678079642" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.074627 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x8zj9" podStartSLOduration=126.074620921 podStartE2EDuration="2m6.074620921s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.030706404 +0000 UTC m=+147.635045009" watchObservedRunningTime="2025-11-22 08:05:26.074620921 +0000 UTC m=+147.678959526" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.089166 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.090626 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.590598937 +0000 UTC m=+148.194937582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.190310 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" podStartSLOduration=125.190290926 podStartE2EDuration="2m5.190290926s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.158789816 +0000 UTC m=+147.763128431" watchObservedRunningTime="2025-11-22 08:05:26.190290926 +0000 UTC m=+147.794629531" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.191616 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzwms" podStartSLOduration=126.191610762 podStartE2EDuration="2m6.191610762s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.190760419 +0000 UTC m=+147.795099024" watchObservedRunningTime="2025-11-22 08:05:26.191610762 +0000 UTC m=+147.795949367" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.191933 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.192319 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.69230417 +0000 UTC m=+148.296642775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.268838 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:26 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:26 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:26 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.268902 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.292951 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.293362 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.793333455 +0000 UTC m=+148.397672060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.394033 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.394343 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.89432952 +0000 UTC m=+148.498668125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.413286 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.413329 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.494726 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.495041 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:26.995016396 +0000 UTC m=+148.599355001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.534393 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.539433 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njhxg" podStartSLOduration=126.539419027 podStartE2EDuration="2m6.539419027s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.538106671 +0000 UTC m=+148.142445276" watchObservedRunningTime="2025-11-22 08:05:26.539419027 +0000 UTC m=+148.143757622" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.540263 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" podStartSLOduration=126.540259069 podStartE2EDuration="2m6.540259069s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.494477371 +0000 UTC m=+148.098815976" watchObservedRunningTime="2025-11-22 08:05:26.540259069 +0000 UTC m=+148.144597674" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.596507 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.596857 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.096844673 +0000 UTC m=+148.701183278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.599217 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-xkl5g" podStartSLOduration=7.599207167 podStartE2EDuration="7.599207167s" podCreationTimestamp="2025-11-22 08:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.570867494 +0000 UTC m=+148.175206099" watchObservedRunningTime="2025-11-22 08:05:26.599207167 +0000 UTC m=+148.203545772" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.642386 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xksr2" podStartSLOduration=125.642373225 podStartE2EDuration="2m5.642373225s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.63850865 +0000 UTC m=+148.242847255" watchObservedRunningTime="2025-11-22 08:05:26.642373225 +0000 UTC m=+148.246711820" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.700677 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.701216 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.201189379 +0000 UTC m=+148.805527984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.802116 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.802746 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.302717357 +0000 UTC m=+148.907055962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.903708 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:26 crc kubenswrapper[4735]: E1122 08:05:26.904078 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.404061831 +0000 UTC m=+149.008400436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.925042 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5j6k5" event={"ID":"e05dba4a-eb90-4ab8-aa71-173e3cd3bfc9","Type":"ContainerStarted","Data":"a88ecf2fcb36872f369ca7fdce954607379f68eb8e191dc03376eb6a241fc656"} Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.927080 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t74q" event={"ID":"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d","Type":"ContainerStarted","Data":"3b4a704bab0d69f43eef0a7ed10b1978b6afb3b525abd310d69c681964eccfee"} Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.928313 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" event={"ID":"a47af7c4-72b1-45bc-95c4-5f3207307fa2","Type":"ContainerStarted","Data":"842654509fadcca5f6afaba2d1dcaeb78f748673071245b8f686977f4f407542"} Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.930332 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" event={"ID":"4b91e784-c670-41af-95c6-696a7197e01a","Type":"ContainerStarted","Data":"78b4eee4a77d74a52c923523700043bee4c56e89e6f72a7cc8a21b0c442c6403"} Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.932801 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5s2w9" event={"ID":"2cf18025-0832-4850-8568-9af2a131711c","Type":"ContainerStarted","Data":"87865af836df504ed698f57d709c4727a8e8446e994132b9d97b6d1af617ddaf"} Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.932828 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.936354 4735 patch_prober.go:28] interesting pod/console-operator-58897d9998-6c285 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.936394 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-6c285" podUID="5dea56bb-e794-479f-8269-e31474c18920" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.936533 4735 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5rsm4 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.936634 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" podUID="2da8d731-f76e-472f-bb05-b2242fb9a652" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.937073 4735 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-v7k7n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.937127 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" podUID="beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.951053 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgp8x" Nov 22 08:05:26 crc kubenswrapper[4735]: I1122 08:05:26.975723 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg8pw" podStartSLOduration=126.975704715 podStartE2EDuration="2m6.975704715s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:26.974286436 +0000 UTC m=+148.578625041" watchObservedRunningTime="2025-11-22 08:05:26.975704715 +0000 UTC m=+148.580043310" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.005917 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.006388 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.506371082 +0000 UTC m=+149.110709677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.068681 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7w2f5" podStartSLOduration=126.06866028 podStartE2EDuration="2m6.06866028s" podCreationTimestamp="2025-11-22 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:27.065813662 +0000 UTC m=+148.670152267" watchObservedRunningTime="2025-11-22 08:05:27.06866028 +0000 UTC m=+148.672998875" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.107636 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.108153 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.608133247 +0000 UTC m=+149.212471852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.212200 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.212615 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.712602445 +0000 UTC m=+149.316941050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.268761 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:27 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:27 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:27 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.268819 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.313532 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.314008 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.813984581 +0000 UTC m=+149.418323186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.382843 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-5s2w9" podStartSLOduration=8.382819398 podStartE2EDuration="8.382819398s" podCreationTimestamp="2025-11-22 08:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:27.133890479 +0000 UTC m=+148.738229084" watchObservedRunningTime="2025-11-22 08:05:27.382819398 +0000 UTC m=+148.987158003" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.385108 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-24jg6"] Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.386246 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.391706 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.405128 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24jg6"] Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.415501 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.415861 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:27.915844339 +0000 UTC m=+149.520182944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.517000 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.517136 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.01711646 +0000 UTC m=+149.621455065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.517279 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-utilities\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.517371 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.517417 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-catalog-content\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.517568 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzk8g\" (UniqueName: \"kubernetes.io/projected/cdab44ae-77cb-454d-b040-c677adfda315-kube-api-access-pzk8g\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.517697 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.017689516 +0000 UTC m=+149.622028121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.582828 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fqrjq"] Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.584510 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.587258 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.609377 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fqrjq"] Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.618932 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.619120 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.119091622 +0000 UTC m=+149.723430227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.619405 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.619457 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-catalog-content\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.619528 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzk8g\" (UniqueName: \"kubernetes.io/projected/cdab44ae-77cb-454d-b040-c677adfda315-kube-api-access-pzk8g\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.619566 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-utilities\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.619792 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.11978132 +0000 UTC m=+149.724119925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.620213 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-utilities\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.620368 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-catalog-content\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.674685 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzk8g\" (UniqueName: \"kubernetes.io/projected/cdab44ae-77cb-454d-b040-c677adfda315-kube-api-access-pzk8g\") pod \"certified-operators-24jg6\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.713316 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.720569 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.720722 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.220698963 +0000 UTC m=+149.825037558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.720806 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-catalog-content\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.720846 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-utilities\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.720909 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.721070 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtp4g\" (UniqueName: \"kubernetes.io/projected/3e5dab20-71fe-425e-b8e0-972f1ac0149a-kube-api-access-xtp4g\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.721332 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.221314159 +0000 UTC m=+149.825652764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.774324 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vmlsp"] Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.775165 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.794792 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vmlsp"] Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.822417 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.822644 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtp4g\" (UniqueName: \"kubernetes.io/projected/3e5dab20-71fe-425e-b8e0-972f1ac0149a-kube-api-access-xtp4g\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.822678 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-catalog-content\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.822700 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-utilities\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.822923 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.32290785 +0000 UTC m=+149.927246455 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.823556 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-catalog-content\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.823765 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-utilities\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.923666 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.923717 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-utilities\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.923755 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-catalog-content\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.923786 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzcrb\" (UniqueName: \"kubernetes.io/projected/c97e92b6-f3cc-4be0-8931-5856d3d83097-kube-api-access-wzcrb\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:27 crc kubenswrapper[4735]: E1122 08:05:27.923975 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.423959215 +0000 UTC m=+150.028297820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.937935 4735 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wmjbb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.938030 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" podUID="e8094c96-1edb-431d-8081-4a4927be57a8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.942945 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtp4g\" (UniqueName: \"kubernetes.io/projected/3e5dab20-71fe-425e-b8e0-972f1ac0149a-kube-api-access-xtp4g\") pod \"community-operators-fqrjq\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:27 crc kubenswrapper[4735]: I1122 08:05:27.985965 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t74q" event={"ID":"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d","Type":"ContainerStarted","Data":"7621ced0ec6058c119e1b731daf6df8b175f0154be7e27f8329999e343f4ba09"} Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.024188 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.024382 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.024408 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-utilities\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.024437 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-catalog-content\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.024496 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzcrb\" (UniqueName: \"kubernetes.io/projected/c97e92b6-f3cc-4be0-8931-5856d3d83097-kube-api-access-wzcrb\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.024542 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.034022 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-catalog-content\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.034246 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-utilities\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.034325 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.534303745 +0000 UTC m=+150.138642350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.039016 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zmzcc"] Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.039342 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.040321 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.043633 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.076148 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.124319 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzcrb\" (UniqueName: \"kubernetes.io/projected/c97e92b6-f3cc-4be0-8931-5856d3d83097-kube-api-access-wzcrb\") pod \"certified-operators-vmlsp\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.125422 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.125449 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d6nc\" (UniqueName: \"kubernetes.io/projected/dd997979-7267-47a2-b486-1b34b36ac6a9-kube-api-access-5d6nc\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.125524 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-utilities\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.125555 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.125572 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.125822 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-catalog-content\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.130046 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.630021186 +0000 UTC m=+150.234359971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.140313 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zmzcc"] Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.144349 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.149261 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.209469 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.227539 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.231110 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d6nc\" (UniqueName: \"kubernetes.io/projected/dd997979-7267-47a2-b486-1b34b36ac6a9-kube-api-access-5d6nc\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.231158 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-utilities\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.231253 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-catalog-content\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.241682 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.74164735 +0000 UTC m=+150.345985955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.242353 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-utilities\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.242660 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-catalog-content\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.282857 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:28 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:28 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:28 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.282931 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.283442 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.319616 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.327769 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.342249 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.342596 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.842577723 +0000 UTC m=+150.446916328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.343279 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d6nc\" (UniqueName: \"kubernetes.io/projected/dd997979-7267-47a2-b486-1b34b36ac6a9-kube-api-access-5d6nc\") pod \"community-operators-zmzcc\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.373654 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.374490 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.377747 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.390736 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.390971 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.391929 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.407702 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24jg6"] Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.419897 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.446067 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.446119 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.946084196 +0000 UTC m=+150.550422801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.446546 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25f2fab7-01fe-485f-9798-cc2793926750-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"25f2fab7-01fe-485f-9798-cc2793926750\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.446608 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.446638 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25f2fab7-01fe-485f-9798-cc2793926750-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"25f2fab7-01fe-485f-9798-cc2793926750\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.447063 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:28.947029831 +0000 UTC m=+150.551368646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.548584 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.549101 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25f2fab7-01fe-485f-9798-cc2793926750-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"25f2fab7-01fe-485f-9798-cc2793926750\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.549181 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25f2fab7-01fe-485f-9798-cc2793926750-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"25f2fab7-01fe-485f-9798-cc2793926750\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.549322 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.04929558 +0000 UTC m=+150.653634185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.551762 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25f2fab7-01fe-485f-9798-cc2793926750-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"25f2fab7-01fe-485f-9798-cc2793926750\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.619023 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25f2fab7-01fe-485f-9798-cc2793926750-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"25f2fab7-01fe-485f-9798-cc2793926750\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.650434 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.650863 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.150835489 +0000 UTC m=+150.755174094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.736134 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.752740 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.753162 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.253137099 +0000 UTC m=+150.857475704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.854511 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.854828 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.354815242 +0000 UTC m=+150.959153847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.956302 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:28 crc kubenswrapper[4735]: E1122 08:05:28.956658 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.456642249 +0000 UTC m=+151.060980854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.995661 4735 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wmjbb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 08:05:28 crc kubenswrapper[4735]: I1122 08:05:28.995721 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" podUID="e8094c96-1edb-431d-8081-4a4927be57a8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.001652 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fqrjq"] Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.058223 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:29 crc kubenswrapper[4735]: E1122 08:05:29.059051 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.559029072 +0000 UTC m=+151.163367667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.089862 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24jg6" event={"ID":"cdab44ae-77cb-454d-b040-c677adfda315","Type":"ContainerStarted","Data":"b27b394e3d579a6d8393b154549e20e7425d84e366001332b7d2754c582ec867"} Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.160650 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:29 crc kubenswrapper[4735]: E1122 08:05:29.163212 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.663191912 +0000 UTC m=+151.267530517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.251063 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zmzcc"] Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.276207 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:29 crc kubenswrapper[4735]: E1122 08:05:29.276508 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.776495652 +0000 UTC m=+151.380834257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9r74q" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.283283 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:29 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:29 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:29 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.283375 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.353576 4735 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.378248 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:29 crc kubenswrapper[4735]: E1122 08:05:29.378936 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 08:05:29.878920796 +0000 UTC m=+151.483259401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.407800 4735 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-22T08:05:29.353596055Z","Handler":null,"Name":""} Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.415083 4735 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.415127 4735 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.507627 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.549405 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.549498 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.620837 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bl92s"] Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.623121 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.634117 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.634610 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl92s"] Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.643025 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vmlsp"] Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.713859 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-utilities\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.713930 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2b4f\" (UniqueName: \"kubernetes.io/projected/331e493d-6996-42ea-84ce-d388ab643352-kube-api-access-k2b4f\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.713992 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-catalog-content\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.732621 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.815431 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-catalog-content\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.815575 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-utilities\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.815638 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2b4f\" (UniqueName: \"kubernetes.io/projected/331e493d-6996-42ea-84ce-d388ab643352-kube-api-access-k2b4f\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.816321 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-catalog-content\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.816550 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-utilities\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.858373 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9r74q\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.877268 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2b4f\" (UniqueName: \"kubernetes.io/projected/331e493d-6996-42ea-84ce-d388ab643352-kube-api-access-k2b4f\") pod \"redhat-marketplace-bl92s\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.918641 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.934803 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.972814 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g58rz"] Nov 22 08:05:29 crc kubenswrapper[4735]: I1122 08:05:29.973741 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.001839 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.013235 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g58rz"] Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.076955 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.134701 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-catalog-content\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.135159 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h96pl\" (UniqueName: \"kubernetes.io/projected/83793793-6201-43a7-85da-9da18c70441f-kube-api-access-h96pl\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.135197 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-utilities\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.155356 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f596da4b15f2728c1d60f4fa839a83a92bdb54faaffd9f396712585d8b0db6a7"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.155413 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"57c431ee7b7d4110c8eec5e5b9937c6e85e1bd294b86a598f5b7d53321064ac8"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.175730 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t74q" event={"ID":"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d","Type":"ContainerStarted","Data":"460be088fc0c5000b97045680bdb4ba5f031abf5f89be8adf75733b5edb0444e"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.175775 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t74q" event={"ID":"fdc81d60-ce15-4ce9-b41c-3815ab3e9b7d","Type":"ContainerStarted","Data":"8c4bfec0ab67b429b4501958ce142aa803cb22fb9eb7217820d414c1f716359d"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.185113 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vmlsp" event={"ID":"c97e92b6-f3cc-4be0-8931-5856d3d83097","Type":"ContainerStarted","Data":"734efa8c79d12cff03bfde2bd3640f0c0b30308c2737882309ec3158c04c9e83"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.185155 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vmlsp" event={"ID":"c97e92b6-f3cc-4be0-8931-5856d3d83097","Type":"ContainerStarted","Data":"2eb4af751f31cdfddb88cc1868a443cbe7a7bfe07ad900e527c5ec1cd7917a47"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.190454 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.201388 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"25f2fab7-01fe-485f-9798-cc2793926750","Type":"ContainerStarted","Data":"233cc624a1fac68f010b67e44a2633d56cf8232b812f8e0fd0a793ffa6fe4537"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.206344 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24jg6" event={"ID":"cdab44ae-77cb-454d-b040-c677adfda315","Type":"ContainerDied","Data":"993efa0fa6d9a52b6e2cc3641d0534abbabba3c41442443e727a333bd6acadff"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.206592 4735 generic.go:334] "Generic (PLEG): container finished" podID="cdab44ae-77cb-454d-b040-c677adfda315" containerID="993efa0fa6d9a52b6e2cc3641d0534abbabba3c41442443e727a333bd6acadff" exitCode=0 Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.225615 4735 generic.go:334] "Generic (PLEG): container finished" podID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerID="7cbff1501ca554f917bd325be6e3d45200555f8979ccf2b95055f38c7a3f092d" exitCode=0 Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.225701 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqrjq" event={"ID":"3e5dab20-71fe-425e-b8e0-972f1ac0149a","Type":"ContainerDied","Data":"7cbff1501ca554f917bd325be6e3d45200555f8979ccf2b95055f38c7a3f092d"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.225735 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqrjq" event={"ID":"3e5dab20-71fe-425e-b8e0-972f1ac0149a","Type":"ContainerStarted","Data":"c6826058da78bb3e0e6020230d4323151edebc2ce5a2bf849f477f23de3face8"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.237995 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-catalog-content\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.238054 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h96pl\" (UniqueName: \"kubernetes.io/projected/83793793-6201-43a7-85da-9da18c70441f-kube-api-access-h96pl\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.238078 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-utilities\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.239335 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-catalog-content\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.239691 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-utilities\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.241162 4735 generic.go:334] "Generic (PLEG): container finished" podID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerID="a632959eb1c747edbf0c2e0f0f86cbcb622853ca3ee726dbd9b151be0868623d" exitCode=0 Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.241258 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmzcc" event={"ID":"dd997979-7267-47a2-b486-1b34b36ac6a9","Type":"ContainerDied","Data":"a632959eb1c747edbf0c2e0f0f86cbcb622853ca3ee726dbd9b151be0868623d"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.241289 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmzcc" event={"ID":"dd997979-7267-47a2-b486-1b34b36ac6a9","Type":"ContainerStarted","Data":"e8dad67d03850cf2eb9d1f3533618fdf9eedcd7af9c84ba10c4b222edec108a9"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.273237 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"bb529b4e565f91ae96c7599c0214240375681a0690df128d81a4ffc5836e78db"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.273279 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7f58dfca97c6084d2249fe34a8dab4214567cac562625ccd09eb627e7a008539"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.274263 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:30 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:30 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:30 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.274317 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.282784 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-4t74q" podStartSLOduration=11.282769195 podStartE2EDuration="11.282769195s" podCreationTimestamp="2025-11-22 08:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:30.279976259 +0000 UTC m=+151.884314864" watchObservedRunningTime="2025-11-22 08:05:30.282769195 +0000 UTC m=+151.887107800" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.290223 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h96pl\" (UniqueName: \"kubernetes.io/projected/83793793-6201-43a7-85da-9da18c70441f-kube-api-access-h96pl\") pod \"redhat-marketplace-g58rz\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.297945 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.311257 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0e9f577c2177a5d3cbb0213dcf3356b594ffd8fdced73b383e2d88530757786a"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.311332 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"aa9031c759c744e394ccd6cbb3a35142aa714054e26891e4879b649b64954068"} Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.312082 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.483414 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9r74q"] Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.513952 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pp9dw" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.562077 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d6s67"] Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.563140 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.570935 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.584896 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d6s67"] Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.618283 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl92s"] Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.662060 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-catalog-content\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.662115 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-utilities\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.662145 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcqqf\" (UniqueName: \"kubernetes.io/projected/f56f24fc-45bb-4842-9de9-2e0f955b04b0-kube-api-access-wcqqf\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.664921 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.670177 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rxw97" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.768140 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-utilities\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.768197 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcqqf\" (UniqueName: \"kubernetes.io/projected/f56f24fc-45bb-4842-9de9-2e0f955b04b0-kube-api-access-wcqqf\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.768300 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-catalog-content\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.769332 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-utilities\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.770231 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-catalog-content\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.801996 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcqqf\" (UniqueName: \"kubernetes.io/projected/f56f24fc-45bb-4842-9de9-2e0f955b04b0-kube-api-access-wcqqf\") pod \"redhat-operators-d6s67\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.909789 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.954729 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f7s4z"] Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.955714 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:30 crc kubenswrapper[4735]: I1122 08:05:30.963343 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f7s4z"] Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.037856 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g58rz"] Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.071730 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-catalog-content\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.071820 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-utilities\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.071887 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t64mq\" (UniqueName: \"kubernetes.io/projected/0e928195-0960-4cef-996e-35cc633b7a86-kube-api-access-t64mq\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: W1122 08:05:31.103946 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83793793_6201_43a7_85da_9da18c70441f.slice/crio-497e578526be7e1fd075fb274db306777fdf34518037e26de118b67ac23e8d82 WatchSource:0}: Error finding container 497e578526be7e1fd075fb274db306777fdf34518037e26de118b67ac23e8d82: Status 404 returned error can't find the container with id 497e578526be7e1fd075fb274db306777fdf34518037e26de118b67ac23e8d82 Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.172983 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-catalog-content\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.173330 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-utilities\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.173353 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t64mq\" (UniqueName: \"kubernetes.io/projected/0e928195-0960-4cef-996e-35cc633b7a86-kube-api-access-t64mq\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.173950 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-catalog-content\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.174011 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-utilities\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.206197 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t64mq\" (UniqueName: \"kubernetes.io/projected/0e928195-0960-4cef-996e-35cc633b7a86-kube-api-access-t64mq\") pod \"redhat-operators-f7s4z\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.268886 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:31 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:31 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:31 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.268938 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.288555 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.291750 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.351479 4735 generic.go:334] "Generic (PLEG): container finished" podID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerID="734efa8c79d12cff03bfde2bd3640f0c0b30308c2737882309ec3158c04c9e83" exitCode=0 Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.351598 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vmlsp" event={"ID":"c97e92b6-f3cc-4be0-8931-5856d3d83097","Type":"ContainerDied","Data":"734efa8c79d12cff03bfde2bd3640f0c0b30308c2737882309ec3158c04c9e83"} Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.361197 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" event={"ID":"bfd8d002-b754-4ee9-842c-48b74cb7ae8f","Type":"ContainerStarted","Data":"f00d1503aad3de00008b1477ea4e0ed604f7d39b7cc113466845254ab2306f69"} Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.361247 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" event={"ID":"bfd8d002-b754-4ee9-842c-48b74cb7ae8f","Type":"ContainerStarted","Data":"29ad8eca9d755b4a7b7d8313f6c43f3ba729c70bc61c2dbde922ef4459dae347"} Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.362117 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.363656 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d6s67"] Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.367060 4735 generic.go:334] "Generic (PLEG): container finished" podID="25f2fab7-01fe-485f-9798-cc2793926750" containerID="28b1044e80760ce5d12eb08eb207e07ed88670d84538de18d2b98355a94ce374" exitCode=0 Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.367197 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"25f2fab7-01fe-485f-9798-cc2793926750","Type":"ContainerDied","Data":"28b1044e80760ce5d12eb08eb207e07ed88670d84538de18d2b98355a94ce374"} Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.369201 4735 generic.go:334] "Generic (PLEG): container finished" podID="331e493d-6996-42ea-84ce-d388ab643352" containerID="5db4815b791297c8c59efb01d3bccbdf2306095f0b249a5f43044ffbf9fb37ed" exitCode=0 Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.369331 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl92s" event={"ID":"331e493d-6996-42ea-84ce-d388ab643352","Type":"ContainerDied","Data":"5db4815b791297c8c59efb01d3bccbdf2306095f0b249a5f43044ffbf9fb37ed"} Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.369358 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl92s" event={"ID":"331e493d-6996-42ea-84ce-d388ab643352","Type":"ContainerStarted","Data":"9f2f213b46e9dd2563085f510c92c20b7827d1f83674e08ac441e4c73912e510"} Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.383593 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" podStartSLOduration=131.383576386 podStartE2EDuration="2m11.383576386s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:31.383508544 +0000 UTC m=+152.987847149" watchObservedRunningTime="2025-11-22 08:05:31.383576386 +0000 UTC m=+152.987914991" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.390520 4735 generic.go:334] "Generic (PLEG): container finished" podID="83793793-6201-43a7-85da-9da18c70441f" containerID="10bd26c2744f3e317ab99bf7d36620e74923b6461ca998371cc3885ce6d169df" exitCode=0 Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.390669 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g58rz" event={"ID":"83793793-6201-43a7-85da-9da18c70441f","Type":"ContainerDied","Data":"10bd26c2744f3e317ab99bf7d36620e74923b6461ca998371cc3885ce6d169df"} Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.390718 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g58rz" event={"ID":"83793793-6201-43a7-85da-9da18c70441f","Type":"ContainerStarted","Data":"497e578526be7e1fd075fb274db306777fdf34518037e26de118b67ac23e8d82"} Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.473659 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-l69gd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.473719 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-l69gd" podUID="048ff883-805e-4a2e-b939-86de1aa2e6e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.474308 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-l69gd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.474326 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-l69gd" podUID="048ff883-805e-4a2e-b939-86de1aa2e6e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.860063 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f7s4z"] Nov 22 08:05:31 crc kubenswrapper[4735]: W1122 08:05:31.904069 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e928195_0960_4cef_996e_35cc633b7a86.slice/crio-7c6b0e5956f758ce83691e8eadeed02f2f62b62131615a36ac2602f3179ff4cc WatchSource:0}: Error finding container 7c6b0e5956f758ce83691e8eadeed02f2f62b62131615a36ac2602f3179ff4cc: Status 404 returned error can't find the container with id 7c6b0e5956f758ce83691e8eadeed02f2f62b62131615a36ac2602f3179ff4cc Nov 22 08:05:31 crc kubenswrapper[4735]: I1122 08:05:31.973057 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-6c285" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.057511 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.057552 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.059042 4735 patch_prober.go:28] interesting pod/console-f9d7485db-sjvm9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.059084 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-sjvm9" podUID="d417a7ee-7687-4086-8047-39e259a9bfbe" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.094834 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.095733 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.101875 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.101883 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.117310 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.200948 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.215552 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1a664c48-2d00-46cc-985e-8066e67fa321-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1a664c48-2d00-46cc-985e-8066e67fa321\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.215639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a664c48-2d00-46cc-985e-8066e67fa321-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1a664c48-2d00-46cc-985e-8066e67fa321\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.262986 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.265450 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:32 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:32 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:32 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.265506 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.326964 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1a664c48-2d00-46cc-985e-8066e67fa321-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1a664c48-2d00-46cc-985e-8066e67fa321\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.327062 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a664c48-2d00-46cc-985e-8066e67fa321-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1a664c48-2d00-46cc-985e-8066e67fa321\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.328582 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1a664c48-2d00-46cc-985e-8066e67fa321-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1a664c48-2d00-46cc-985e-8066e67fa321\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.360517 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a664c48-2d00-46cc-985e-8066e67fa321-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1a664c48-2d00-46cc-985e-8066e67fa321\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.429555 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.455709 4735 generic.go:334] "Generic (PLEG): container finished" podID="0e928195-0960-4cef-996e-35cc633b7a86" containerID="49ea8b71b387b9646b41aedda6b3a72a9eed818abc024568302e89d919c63aa4" exitCode=0 Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.455781 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f7s4z" event={"ID":"0e928195-0960-4cef-996e-35cc633b7a86","Type":"ContainerDied","Data":"49ea8b71b387b9646b41aedda6b3a72a9eed818abc024568302e89d919c63aa4"} Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.455811 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f7s4z" event={"ID":"0e928195-0960-4cef-996e-35cc633b7a86","Type":"ContainerStarted","Data":"7c6b0e5956f758ce83691e8eadeed02f2f62b62131615a36ac2602f3179ff4cc"} Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.463851 4735 generic.go:334] "Generic (PLEG): container finished" podID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerID="033bee9303354f6215d8e67cb5549e7322d871a760cf20faab81bcf715fa2b7c" exitCode=0 Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.464762 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6s67" event={"ID":"f56f24fc-45bb-4842-9de9-2e0f955b04b0","Type":"ContainerDied","Data":"033bee9303354f6215d8e67cb5549e7322d871a760cf20faab81bcf715fa2b7c"} Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.464820 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6s67" event={"ID":"f56f24fc-45bb-4842-9de9-2e0f955b04b0","Type":"ContainerStarted","Data":"de73a2ad05299894354afdad0d995eb3cbe6be667e3bceae589f430575078f1c"} Nov 22 08:05:32 crc kubenswrapper[4735]: I1122 08:05:32.496525 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wmjbb" Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.004368 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.087991 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.156988 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25f2fab7-01fe-485f-9798-cc2793926750-kubelet-dir\") pod \"25f2fab7-01fe-485f-9798-cc2793926750\" (UID: \"25f2fab7-01fe-485f-9798-cc2793926750\") " Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.157053 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25f2fab7-01fe-485f-9798-cc2793926750-kube-api-access\") pod \"25f2fab7-01fe-485f-9798-cc2793926750\" (UID: \"25f2fab7-01fe-485f-9798-cc2793926750\") " Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.158141 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25f2fab7-01fe-485f-9798-cc2793926750-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "25f2fab7-01fe-485f-9798-cc2793926750" (UID: "25f2fab7-01fe-485f-9798-cc2793926750"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.178838 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f2fab7-01fe-485f-9798-cc2793926750-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "25f2fab7-01fe-485f-9798-cc2793926750" (UID: "25f2fab7-01fe-485f-9798-cc2793926750"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.258303 4735 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25f2fab7-01fe-485f-9798-cc2793926750-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.258336 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25f2fab7-01fe-485f-9798-cc2793926750-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.269608 4735 patch_prober.go:28] interesting pod/router-default-5444994796-w5v4q container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 08:05:33 crc kubenswrapper[4735]: [-]has-synced failed: reason withheld Nov 22 08:05:33 crc kubenswrapper[4735]: [+]process-running ok Nov 22 08:05:33 crc kubenswrapper[4735]: healthz check failed Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.269667 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5v4q" podUID="53cd51d1-1576-492a-ba5e-e26a388b6a57" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.485842 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1a664c48-2d00-46cc-985e-8066e67fa321","Type":"ContainerStarted","Data":"231587bb279f31c8a82cd275df555fbe6f2c032b5a791c1f754e36ca7678a8b7"} Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.529560 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"25f2fab7-01fe-485f-9798-cc2793926750","Type":"ContainerDied","Data":"233cc624a1fac68f010b67e44a2633d56cf8232b812f8e0fd0a793ffa6fe4537"} Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.529597 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="233cc624a1fac68f010b67e44a2633d56cf8232b812f8e0fd0a793ffa6fe4537" Nov 22 08:05:33 crc kubenswrapper[4735]: I1122 08:05:33.529656 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 08:05:34 crc kubenswrapper[4735]: I1122 08:05:34.265307 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:34 crc kubenswrapper[4735]: I1122 08:05:34.273718 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-w5v4q" Nov 22 08:05:34 crc kubenswrapper[4735]: I1122 08:05:34.598302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1a664c48-2d00-46cc-985e-8066e67fa321","Type":"ContainerStarted","Data":"3be1c9a28f2b4667da3a1917a1224cc9c3e7a3835acf5a63a3f8d00a16e62126"} Nov 22 08:05:34 crc kubenswrapper[4735]: I1122 08:05:34.621431 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.621404928 podStartE2EDuration="2.621404928s" podCreationTimestamp="2025-11-22 08:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:05:34.610357857 +0000 UTC m=+156.214696472" watchObservedRunningTime="2025-11-22 08:05:34.621404928 +0000 UTC m=+156.225743533" Nov 22 08:05:35 crc kubenswrapper[4735]: I1122 08:05:35.614597 4735 generic.go:334] "Generic (PLEG): container finished" podID="1a664c48-2d00-46cc-985e-8066e67fa321" containerID="3be1c9a28f2b4667da3a1917a1224cc9c3e7a3835acf5a63a3f8d00a16e62126" exitCode=0 Nov 22 08:05:35 crc kubenswrapper[4735]: I1122 08:05:35.614699 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1a664c48-2d00-46cc-985e-8066e67fa321","Type":"ContainerDied","Data":"3be1c9a28f2b4667da3a1917a1224cc9c3e7a3835acf5a63a3f8d00a16e62126"} Nov 22 08:05:35 crc kubenswrapper[4735]: I1122 08:05:35.622486 4735 generic.go:334] "Generic (PLEG): container finished" podID="31a557ae-6002-49fd-a9af-b3ea23dfd952" containerID="edf114b0a6ecb80547997ca80f1020152ca9adb4047e825a975d92c607a1f28c" exitCode=0 Nov 22 08:05:35 crc kubenswrapper[4735]: I1122 08:05:35.622538 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" event={"ID":"31a557ae-6002-49fd-a9af-b3ea23dfd952","Type":"ContainerDied","Data":"edf114b0a6ecb80547997ca80f1020152ca9adb4047e825a975d92c607a1f28c"} Nov 22 08:05:36 crc kubenswrapper[4735]: I1122 08:05:36.975427 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.017037 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.142510 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31a557ae-6002-49fd-a9af-b3ea23dfd952-config-volume\") pod \"31a557ae-6002-49fd-a9af-b3ea23dfd952\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.142549 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1a664c48-2d00-46cc-985e-8066e67fa321-kubelet-dir\") pod \"1a664c48-2d00-46cc-985e-8066e67fa321\" (UID: \"1a664c48-2d00-46cc-985e-8066e67fa321\") " Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.142573 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc9cf\" (UniqueName: \"kubernetes.io/projected/31a557ae-6002-49fd-a9af-b3ea23dfd952-kube-api-access-sc9cf\") pod \"31a557ae-6002-49fd-a9af-b3ea23dfd952\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.142702 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31a557ae-6002-49fd-a9af-b3ea23dfd952-secret-volume\") pod \"31a557ae-6002-49fd-a9af-b3ea23dfd952\" (UID: \"31a557ae-6002-49fd-a9af-b3ea23dfd952\") " Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.142732 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a664c48-2d00-46cc-985e-8066e67fa321-kube-api-access\") pod \"1a664c48-2d00-46cc-985e-8066e67fa321\" (UID: \"1a664c48-2d00-46cc-985e-8066e67fa321\") " Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.149638 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a664c48-2d00-46cc-985e-8066e67fa321-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1a664c48-2d00-46cc-985e-8066e67fa321" (UID: "1a664c48-2d00-46cc-985e-8066e67fa321"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.150956 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31a557ae-6002-49fd-a9af-b3ea23dfd952-config-volume" (OuterVolumeSpecName: "config-volume") pod "31a557ae-6002-49fd-a9af-b3ea23dfd952" (UID: "31a557ae-6002-49fd-a9af-b3ea23dfd952"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.170316 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31a557ae-6002-49fd-a9af-b3ea23dfd952-kube-api-access-sc9cf" (OuterVolumeSpecName: "kube-api-access-sc9cf") pod "31a557ae-6002-49fd-a9af-b3ea23dfd952" (UID: "31a557ae-6002-49fd-a9af-b3ea23dfd952"). InnerVolumeSpecName "kube-api-access-sc9cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.170905 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31a557ae-6002-49fd-a9af-b3ea23dfd952-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "31a557ae-6002-49fd-a9af-b3ea23dfd952" (UID: "31a557ae-6002-49fd-a9af-b3ea23dfd952"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.175724 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a664c48-2d00-46cc-985e-8066e67fa321-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1a664c48-2d00-46cc-985e-8066e67fa321" (UID: "1a664c48-2d00-46cc-985e-8066e67fa321"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.244325 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a664c48-2d00-46cc-985e-8066e67fa321-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.244371 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31a557ae-6002-49fd-a9af-b3ea23dfd952-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.244385 4735 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1a664c48-2d00-46cc-985e-8066e67fa321-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.244397 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc9cf\" (UniqueName: \"kubernetes.io/projected/31a557ae-6002-49fd-a9af-b3ea23dfd952-kube-api-access-sc9cf\") on node \"crc\" DevicePath \"\"" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.244411 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31a557ae-6002-49fd-a9af-b3ea23dfd952-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.333064 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-5s2w9" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.644891 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1a664c48-2d00-46cc-985e-8066e67fa321","Type":"ContainerDied","Data":"231587bb279f31c8a82cd275df555fbe6f2c032b5a791c1f754e36ca7678a8b7"} Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.645163 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="231587bb279f31c8a82cd275df555fbe6f2c032b5a791c1f754e36ca7678a8b7" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.644940 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.652033 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" event={"ID":"31a557ae-6002-49fd-a9af-b3ea23dfd952","Type":"ContainerDied","Data":"a3577f7bfba6b019d9c0ee4100f9e668ca6567238bd3ded96c2a5136123a2d01"} Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.652078 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3577f7bfba6b019d9c0ee4100f9e668ca6567238bd3ded96c2a5136123a2d01" Nov 22 08:05:37 crc kubenswrapper[4735]: I1122 08:05:37.652106 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q" Nov 22 08:05:41 crc kubenswrapper[4735]: I1122 08:05:41.470539 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-l69gd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 22 08:05:41 crc kubenswrapper[4735]: I1122 08:05:41.470532 4735 patch_prober.go:28] interesting pod/downloads-7954f5f757-l69gd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 22 08:05:41 crc kubenswrapper[4735]: I1122 08:05:41.471074 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-l69gd" podUID="048ff883-805e-4a2e-b939-86de1aa2e6e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 22 08:05:41 crc kubenswrapper[4735]: I1122 08:05:41.471130 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-l69gd" podUID="048ff883-805e-4a2e-b939-86de1aa2e6e2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 22 08:05:42 crc kubenswrapper[4735]: I1122 08:05:42.061300 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:42 crc kubenswrapper[4735]: I1122 08:05:42.066159 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:05:43 crc kubenswrapper[4735]: I1122 08:05:43.226164 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:43 crc kubenswrapper[4735]: I1122 08:05:43.232115 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9170f64-5b0d-4b2b-99ce-7ecb9f567620-metrics-certs\") pod \"network-metrics-daemon-vvf57\" (UID: \"e9170f64-5b0d-4b2b-99ce-7ecb9f567620\") " pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:43 crc kubenswrapper[4735]: I1122 08:05:43.486585 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vvf57" Nov 22 08:05:46 crc kubenswrapper[4735]: I1122 08:05:46.132048 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:05:46 crc kubenswrapper[4735]: I1122 08:05:46.132659 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:05:49 crc kubenswrapper[4735]: I1122 08:05:49.941504 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:05:51 crc kubenswrapper[4735]: I1122 08:05:51.486504 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-l69gd" Nov 22 08:06:02 crc kubenswrapper[4735]: I1122 08:06:02.225123 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g5sst" Nov 22 08:06:05 crc kubenswrapper[4735]: E1122 08:06:05.583817 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 22 08:06:05 crc kubenswrapper[4735]: E1122 08:06:05.584290 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pzk8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-24jg6_openshift-marketplace(cdab44ae-77cb-454d-b040-c677adfda315): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 08:06:05 crc kubenswrapper[4735]: E1122 08:06:05.585493 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-24jg6" podUID="cdab44ae-77cb-454d-b040-c677adfda315" Nov 22 08:06:08 crc kubenswrapper[4735]: I1122 08:06:08.367291 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 08:06:10 crc kubenswrapper[4735]: E1122 08:06:10.269851 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-24jg6" podUID="cdab44ae-77cb-454d-b040-c677adfda315" Nov 22 08:06:11 crc kubenswrapper[4735]: E1122 08:06:11.776153 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 22 08:06:11 crc kubenswrapper[4735]: E1122 08:06:11.776359 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wcqqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-d6s67_openshift-marketplace(f56f24fc-45bb-4842-9de9-2e0f955b04b0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 08:06:11 crc kubenswrapper[4735]: E1122 08:06:11.777616 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-d6s67" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" Nov 22 08:06:14 crc kubenswrapper[4735]: E1122 08:06:14.597129 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-d6s67" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" Nov 22 08:06:14 crc kubenswrapper[4735]: E1122 08:06:14.710338 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 22 08:06:14 crc kubenswrapper[4735]: E1122 08:06:14.710808 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k2b4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bl92s_openshift-marketplace(331e493d-6996-42ea-84ce-d388ab643352): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 08:06:14 crc kubenswrapper[4735]: E1122 08:06:14.711760 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 22 08:06:14 crc kubenswrapper[4735]: E1122 08:06:14.711917 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h96pl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-g58rz_openshift-marketplace(83793793-6201-43a7-85da-9da18c70441f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 08:06:14 crc kubenswrapper[4735]: E1122 08:06:14.711976 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bl92s" podUID="331e493d-6996-42ea-84ce-d388ab643352" Nov 22 08:06:14 crc kubenswrapper[4735]: E1122 08:06:14.713411 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-g58rz" podUID="83793793-6201-43a7-85da-9da18c70441f" Nov 22 08:06:16 crc kubenswrapper[4735]: I1122 08:06:16.131811 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:06:16 crc kubenswrapper[4735]: I1122 08:06:16.131890 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.321190 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bl92s" podUID="331e493d-6996-42ea-84ce-d388ab643352" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.321261 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-g58rz" podUID="83793793-6201-43a7-85da-9da18c70441f" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.425913 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.426306 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5d6nc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zmzcc_openshift-marketplace(dd997979-7267-47a2-b486-1b34b36ac6a9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.428351 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zmzcc" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.450851 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.451001 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t64mq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-f7s4z_openshift-marketplace(0e928195-0960-4cef-996e-35cc633b7a86): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.451929 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.452047 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xtp4g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-fqrjq_openshift-marketplace(3e5dab20-71fe-425e-b8e0-972f1ac0149a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.452100 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-f7s4z" podUID="0e928195-0960-4cef-996e-35cc633b7a86" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.453272 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-fqrjq" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.509766 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.509952 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wzcrb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vmlsp_openshift-marketplace(c97e92b6-f3cc-4be0-8931-5856d3d83097): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.511920 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vmlsp" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" Nov 22 08:06:17 crc kubenswrapper[4735]: I1122 08:06:17.795781 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vvf57"] Nov 22 08:06:17 crc kubenswrapper[4735]: W1122 08:06:17.803200 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9170f64_5b0d_4b2b_99ce_7ecb9f567620.slice/crio-f5a89538587776e5999d15b75de5e131809445172aee90feb03f77b148ac41f6 WatchSource:0}: Error finding container f5a89538587776e5999d15b75de5e131809445172aee90feb03f77b148ac41f6: Status 404 returned error can't find the container with id f5a89538587776e5999d15b75de5e131809445172aee90feb03f77b148ac41f6 Nov 22 08:06:17 crc kubenswrapper[4735]: I1122 08:06:17.884530 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vvf57" event={"ID":"e9170f64-5b0d-4b2b-99ce-7ecb9f567620","Type":"ContainerStarted","Data":"f5a89538587776e5999d15b75de5e131809445172aee90feb03f77b148ac41f6"} Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.886094 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zmzcc" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.886933 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-fqrjq" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.887324 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-f7s4z" podUID="0e928195-0960-4cef-996e-35cc633b7a86" Nov 22 08:06:17 crc kubenswrapper[4735]: E1122 08:06:17.888888 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vmlsp" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" Nov 22 08:06:18 crc kubenswrapper[4735]: I1122 08:06:18.892409 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vvf57" event={"ID":"e9170f64-5b0d-4b2b-99ce-7ecb9f567620","Type":"ContainerStarted","Data":"b134eb3ea2ebc82047b4ebe583977041b7b4895a2f64f6b582bdba01f0b3f6eb"} Nov 22 08:06:18 crc kubenswrapper[4735]: I1122 08:06:18.892739 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vvf57" event={"ID":"e9170f64-5b0d-4b2b-99ce-7ecb9f567620","Type":"ContainerStarted","Data":"e374c330e7f171cbb3cd87c40a2639e409e859db9f49d1819e20ecade42da670"} Nov 22 08:06:18 crc kubenswrapper[4735]: I1122 08:06:18.913244 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vvf57" podStartSLOduration=178.913213288 podStartE2EDuration="2m58.913213288s" podCreationTimestamp="2025-11-22 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:06:18.908996173 +0000 UTC m=+200.513334778" watchObservedRunningTime="2025-11-22 08:06:18.913213288 +0000 UTC m=+200.517551933" Nov 22 08:06:25 crc kubenswrapper[4735]: I1122 08:06:25.939055 4735 generic.go:334] "Generic (PLEG): container finished" podID="cdab44ae-77cb-454d-b040-c677adfda315" containerID="b283e7e776d674d16413252df4861bbc2860eee2c68ba0d765a7b3959632fb77" exitCode=0 Nov 22 08:06:25 crc kubenswrapper[4735]: I1122 08:06:25.939109 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24jg6" event={"ID":"cdab44ae-77cb-454d-b040-c677adfda315","Type":"ContainerDied","Data":"b283e7e776d674d16413252df4861bbc2860eee2c68ba0d765a7b3959632fb77"} Nov 22 08:06:26 crc kubenswrapper[4735]: I1122 08:06:26.946315 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24jg6" event={"ID":"cdab44ae-77cb-454d-b040-c677adfda315","Type":"ContainerStarted","Data":"ff584f951c30542bb3daaaa779121aa167a93c9a8ae1580349955fcdd31e9105"} Nov 22 08:06:26 crc kubenswrapper[4735]: I1122 08:06:26.960348 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-24jg6" podStartSLOduration=3.404508157 podStartE2EDuration="59.960331259s" podCreationTimestamp="2025-11-22 08:05:27 +0000 UTC" firstStartedPulling="2025-11-22 08:05:30.208442438 +0000 UTC m=+151.812781043" lastFinishedPulling="2025-11-22 08:06:26.76426555 +0000 UTC m=+208.368604145" observedRunningTime="2025-11-22 08:06:26.960081532 +0000 UTC m=+208.564420137" watchObservedRunningTime="2025-11-22 08:06:26.960331259 +0000 UTC m=+208.564669854" Nov 22 08:06:27 crc kubenswrapper[4735]: I1122 08:06:27.713799 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:06:27 crc kubenswrapper[4735]: I1122 08:06:27.713866 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:06:28 crc kubenswrapper[4735]: I1122 08:06:28.846440 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-24jg6" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="registry-server" probeResult="failure" output=< Nov 22 08:06:28 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:06:28 crc kubenswrapper[4735]: > Nov 22 08:06:29 crc kubenswrapper[4735]: I1122 08:06:29.962572 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6s67" event={"ID":"f56f24fc-45bb-4842-9de9-2e0f955b04b0","Type":"ContainerStarted","Data":"aaee914902b1945e64e472ac5b0606fb6291dc61c1fe4aaba33176db07cee75d"} Nov 22 08:06:30 crc kubenswrapper[4735]: I1122 08:06:30.969355 4735 generic.go:334] "Generic (PLEG): container finished" podID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerID="aaee914902b1945e64e472ac5b0606fb6291dc61c1fe4aaba33176db07cee75d" exitCode=0 Nov 22 08:06:30 crc kubenswrapper[4735]: I1122 08:06:30.969519 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6s67" event={"ID":"f56f24fc-45bb-4842-9de9-2e0f955b04b0","Type":"ContainerDied","Data":"aaee914902b1945e64e472ac5b0606fb6291dc61c1fe4aaba33176db07cee75d"} Nov 22 08:06:31 crc kubenswrapper[4735]: I1122 08:06:31.989564 4735 generic.go:334] "Generic (PLEG): container finished" podID="0e928195-0960-4cef-996e-35cc633b7a86" containerID="c1337b61a02682a1ac56eb2d5236e6b563d07aec58608852aa4364e23c692637" exitCode=0 Nov 22 08:06:31 crc kubenswrapper[4735]: I1122 08:06:31.989657 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f7s4z" event={"ID":"0e928195-0960-4cef-996e-35cc633b7a86","Type":"ContainerDied","Data":"c1337b61a02682a1ac56eb2d5236e6b563d07aec58608852aa4364e23c692637"} Nov 22 08:06:31 crc kubenswrapper[4735]: I1122 08:06:31.993902 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6s67" event={"ID":"f56f24fc-45bb-4842-9de9-2e0f955b04b0","Type":"ContainerStarted","Data":"f33c76a4ae3a9f1004dba5a492066597b1df427f03d1c7741ed2e1091b44b611"} Nov 22 08:06:31 crc kubenswrapper[4735]: I1122 08:06:31.997987 4735 generic.go:334] "Generic (PLEG): container finished" podID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerID="6eae83fae1169b46961a643ee4bf4d38a5e5766dad95d8669eae379262e59b45" exitCode=0 Nov 22 08:06:31 crc kubenswrapper[4735]: I1122 08:06:31.998022 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqrjq" event={"ID":"3e5dab20-71fe-425e-b8e0-972f1ac0149a","Type":"ContainerDied","Data":"6eae83fae1169b46961a643ee4bf4d38a5e5766dad95d8669eae379262e59b45"} Nov 22 08:06:32 crc kubenswrapper[4735]: I1122 08:06:32.288437 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d6s67" podStartSLOduration=3.38973789 podStartE2EDuration="1m2.288417434s" podCreationTimestamp="2025-11-22 08:05:30 +0000 UTC" firstStartedPulling="2025-11-22 08:05:32.48666152 +0000 UTC m=+154.091000125" lastFinishedPulling="2025-11-22 08:06:31.385341064 +0000 UTC m=+212.989679669" observedRunningTime="2025-11-22 08:06:32.090437946 +0000 UTC m=+213.694776551" watchObservedRunningTime="2025-11-22 08:06:32.288417434 +0000 UTC m=+213.892756029" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.004942 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vmlsp" event={"ID":"c97e92b6-f3cc-4be0-8931-5856d3d83097","Type":"ContainerStarted","Data":"a79925bc3ff4d047d4d754580e520d760a920c6f26ad09f9298c9950a5ec636e"} Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.007440 4735 generic.go:334] "Generic (PLEG): container finished" podID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerID="ad459ac5d0f4dc394efbb2624eaf4ba9037b13470b81c82f39b3f11451605aca" exitCode=0 Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.007466 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmzcc" event={"ID":"dd997979-7267-47a2-b486-1b34b36ac6a9","Type":"ContainerDied","Data":"ad459ac5d0f4dc394efbb2624eaf4ba9037b13470b81c82f39b3f11451605aca"} Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.009544 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl92s" event={"ID":"331e493d-6996-42ea-84ce-d388ab643352","Type":"ContainerStarted","Data":"16873a97676e311f9b97c000d1dd788b7dd996b293d1de9da5a638a9be08f9ed"} Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.011374 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f7s4z" event={"ID":"0e928195-0960-4cef-996e-35cc633b7a86","Type":"ContainerStarted","Data":"1092958533a50e85699c2f320db970bf10c84f45745a11fa07fe4fb65a4cf586"} Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.013394 4735 generic.go:334] "Generic (PLEG): container finished" podID="83793793-6201-43a7-85da-9da18c70441f" containerID="5aa828dd9baa214f29a92743a91485338e102ff2d6267349dabe165da7ef3f0e" exitCode=0 Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.013447 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g58rz" event={"ID":"83793793-6201-43a7-85da-9da18c70441f","Type":"ContainerDied","Data":"5aa828dd9baa214f29a92743a91485338e102ff2d6267349dabe165da7ef3f0e"} Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.018509 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqrjq" event={"ID":"3e5dab20-71fe-425e-b8e0-972f1ac0149a","Type":"ContainerStarted","Data":"b2ce3e5ede47208f510fd30065bd8cb855d944474eefe9b12794d378a6cf9bb5"} Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.083799 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fqrjq" podStartSLOduration=3.823382812 podStartE2EDuration="1m6.083784779s" podCreationTimestamp="2025-11-22 08:05:27 +0000 UTC" firstStartedPulling="2025-11-22 08:05:30.227582821 +0000 UTC m=+151.831921426" lastFinishedPulling="2025-11-22 08:06:32.487984788 +0000 UTC m=+214.092323393" observedRunningTime="2025-11-22 08:06:33.083551563 +0000 UTC m=+214.687890168" watchObservedRunningTime="2025-11-22 08:06:33.083784779 +0000 UTC m=+214.688123384" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.704275 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f7s4z" podStartSLOduration=3.7561959160000002 podStartE2EDuration="1m3.704257117s" podCreationTimestamp="2025-11-22 08:05:30 +0000 UTC" firstStartedPulling="2025-11-22 08:05:32.48665339 +0000 UTC m=+154.090991985" lastFinishedPulling="2025-11-22 08:06:32.434714581 +0000 UTC m=+214.039053186" observedRunningTime="2025-11-22 08:06:33.123386367 +0000 UTC m=+214.727724992" watchObservedRunningTime="2025-11-22 08:06:33.704257117 +0000 UTC m=+215.308595722" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.705085 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wmm22"] Nov 22 08:06:33 crc kubenswrapper[4735]: E1122 08:06:33.705420 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a664c48-2d00-46cc-985e-8066e67fa321" containerName="pruner" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.705578 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a664c48-2d00-46cc-985e-8066e67fa321" containerName="pruner" Nov 22 08:06:33 crc kubenswrapper[4735]: E1122 08:06:33.705649 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a557ae-6002-49fd-a9af-b3ea23dfd952" containerName="collect-profiles" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.705719 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a557ae-6002-49fd-a9af-b3ea23dfd952" containerName="collect-profiles" Nov 22 08:06:33 crc kubenswrapper[4735]: E1122 08:06:33.705783 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f2fab7-01fe-485f-9798-cc2793926750" containerName="pruner" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.705845 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f2fab7-01fe-485f-9798-cc2793926750" containerName="pruner" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.705993 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a557ae-6002-49fd-a9af-b3ea23dfd952" containerName="collect-profiles" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.706063 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="25f2fab7-01fe-485f-9798-cc2793926750" containerName="pruner" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.706125 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a664c48-2d00-46cc-985e-8066e67fa321" containerName="pruner" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.706597 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.727305 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wmm22"] Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.841127 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.841183 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-registry-tls\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.841200 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-trusted-ca\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.841238 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.841256 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrmh\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-kube-api-access-5xrmh\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.841272 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-bound-sa-token\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.841289 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-registry-certificates\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.841320 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.876703 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.942219 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.942276 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-registry-tls\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.942295 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-trusted-ca\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.942328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-bound-sa-token\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.942344 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrmh\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-kube-api-access-5xrmh\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.942361 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-registry-certificates\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.942388 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.942896 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.943962 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-registry-certificates\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.944759 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-trusted-ca\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.961246 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.964985 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-registry-tls\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.965686 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-bound-sa-token\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:33 crc kubenswrapper[4735]: I1122 08:06:33.965715 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrmh\" (UniqueName: \"kubernetes.io/projected/1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7-kube-api-access-5xrmh\") pod \"image-registry-66df7c8f76-wmm22\" (UID: \"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7\") " pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.019538 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.026743 4735 generic.go:334] "Generic (PLEG): container finished" podID="331e493d-6996-42ea-84ce-d388ab643352" containerID="16873a97676e311f9b97c000d1dd788b7dd996b293d1de9da5a638a9be08f9ed" exitCode=0 Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.026837 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl92s" event={"ID":"331e493d-6996-42ea-84ce-d388ab643352","Type":"ContainerDied","Data":"16873a97676e311f9b97c000d1dd788b7dd996b293d1de9da5a638a9be08f9ed"} Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.028880 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g58rz" event={"ID":"83793793-6201-43a7-85da-9da18c70441f","Type":"ContainerStarted","Data":"8e5fae32fb0b7b9348272aacd905bc66b744fbd847d44795a7f89a638431813a"} Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.030950 4735 generic.go:334] "Generic (PLEG): container finished" podID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerID="a79925bc3ff4d047d4d754580e520d760a920c6f26ad09f9298c9950a5ec636e" exitCode=0 Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.031019 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vmlsp" event={"ID":"c97e92b6-f3cc-4be0-8931-5856d3d83097","Type":"ContainerDied","Data":"a79925bc3ff4d047d4d754580e520d760a920c6f26ad09f9298c9950a5ec636e"} Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.037498 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmzcc" event={"ID":"dd997979-7267-47a2-b486-1b34b36ac6a9","Type":"ContainerStarted","Data":"9aeada7369d1f7a19d4885baec8c1ce94eca5fa3f8f397501ae5592137fa640d"} Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.095585 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zmzcc" podStartSLOduration=3.941557299 podStartE2EDuration="1m7.095567252s" podCreationTimestamp="2025-11-22 08:05:27 +0000 UTC" firstStartedPulling="2025-11-22 08:05:30.261643459 +0000 UTC m=+151.865982064" lastFinishedPulling="2025-11-22 08:06:33.415653412 +0000 UTC m=+215.019992017" observedRunningTime="2025-11-22 08:06:34.092503669 +0000 UTC m=+215.696842274" watchObservedRunningTime="2025-11-22 08:06:34.095567252 +0000 UTC m=+215.699905857" Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.108777 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g58rz" podStartSLOduration=3.121426893 podStartE2EDuration="1m5.108762518s" podCreationTimestamp="2025-11-22 08:05:29 +0000 UTC" firstStartedPulling="2025-11-22 08:05:31.392907721 +0000 UTC m=+152.997246326" lastFinishedPulling="2025-11-22 08:06:33.380243346 +0000 UTC m=+214.984581951" observedRunningTime="2025-11-22 08:06:34.106265371 +0000 UTC m=+215.710603976" watchObservedRunningTime="2025-11-22 08:06:34.108762518 +0000 UTC m=+215.713101123" Nov 22 08:06:34 crc kubenswrapper[4735]: I1122 08:06:34.585125 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wmm22"] Nov 22 08:06:34 crc kubenswrapper[4735]: W1122 08:06:34.595165 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cb9e8d0_5cf5_48ab_a7d5_831a9afe1dd7.slice/crio-0966e9fa7ad9a5bd8da61d051a7cf2c252d6637ca94f7d5822e5f78718386613 WatchSource:0}: Error finding container 0966e9fa7ad9a5bd8da61d051a7cf2c252d6637ca94f7d5822e5f78718386613: Status 404 returned error can't find the container with id 0966e9fa7ad9a5bd8da61d051a7cf2c252d6637ca94f7d5822e5f78718386613 Nov 22 08:06:35 crc kubenswrapper[4735]: I1122 08:06:35.044344 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" event={"ID":"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7","Type":"ContainerStarted","Data":"450ba4777bf904ca6c00819ceea3aa69664cdd4572b128fb1585ace88dc48fdc"} Nov 22 08:06:35 crc kubenswrapper[4735]: I1122 08:06:35.044585 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" event={"ID":"1cb9e8d0-5cf5-48ab-a7d5-831a9afe1dd7","Type":"ContainerStarted","Data":"0966e9fa7ad9a5bd8da61d051a7cf2c252d6637ca94f7d5822e5f78718386613"} Nov 22 08:06:35 crc kubenswrapper[4735]: I1122 08:06:35.044603 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:35 crc kubenswrapper[4735]: I1122 08:06:35.047080 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vmlsp" event={"ID":"c97e92b6-f3cc-4be0-8931-5856d3d83097","Type":"ContainerStarted","Data":"514893f8c26757a4d29715d50854383eed3b183826d3c44ceeccda96219694b9"} Nov 22 08:06:35 crc kubenswrapper[4735]: I1122 08:06:35.049769 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl92s" event={"ID":"331e493d-6996-42ea-84ce-d388ab643352","Type":"ContainerStarted","Data":"111109e83efbee8881aa414acc422c0bb3fd638f020b5ad28a80491e10da537d"} Nov 22 08:06:35 crc kubenswrapper[4735]: I1122 08:06:35.070662 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" podStartSLOduration=2.070645225 podStartE2EDuration="2.070645225s" podCreationTimestamp="2025-11-22 08:06:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:06:35.065554068 +0000 UTC m=+216.669892673" watchObservedRunningTime="2025-11-22 08:06:35.070645225 +0000 UTC m=+216.674983830" Nov 22 08:06:35 crc kubenswrapper[4735]: I1122 08:06:35.082553 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vmlsp" podStartSLOduration=3.75035073 podStartE2EDuration="1m8.082532226s" podCreationTimestamp="2025-11-22 08:05:27 +0000 UTC" firstStartedPulling="2025-11-22 08:05:30.1901993 +0000 UTC m=+151.794537905" lastFinishedPulling="2025-11-22 08:06:34.522380796 +0000 UTC m=+216.126719401" observedRunningTime="2025-11-22 08:06:35.081539419 +0000 UTC m=+216.685878034" watchObservedRunningTime="2025-11-22 08:06:35.082532226 +0000 UTC m=+216.686870841" Nov 22 08:06:35 crc kubenswrapper[4735]: I1122 08:06:35.103148 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bl92s" podStartSLOduration=3.005536133 podStartE2EDuration="1m6.103129231s" podCreationTimestamp="2025-11-22 08:05:29 +0000 UTC" firstStartedPulling="2025-11-22 08:05:31.38592586 +0000 UTC m=+152.990264465" lastFinishedPulling="2025-11-22 08:06:34.483518958 +0000 UTC m=+216.087857563" observedRunningTime="2025-11-22 08:06:35.099422902 +0000 UTC m=+216.703761507" watchObservedRunningTime="2025-11-22 08:06:35.103129231 +0000 UTC m=+216.707467836" Nov 22 08:06:37 crc kubenswrapper[4735]: I1122 08:06:37.846853 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:06:37 crc kubenswrapper[4735]: I1122 08:06:37.899425 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.210767 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.211046 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.254248 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.392834 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.393114 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.420825 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.420896 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.443827 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:06:38 crc kubenswrapper[4735]: I1122 08:06:38.455750 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:06:39 crc kubenswrapper[4735]: I1122 08:06:39.109846 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:06:39 crc kubenswrapper[4735]: I1122 08:06:39.110857 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:06:39 crc kubenswrapper[4735]: I1122 08:06:39.115214 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:06:39 crc kubenswrapper[4735]: I1122 08:06:39.971568 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zmzcc"] Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.078358 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.078401 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.130235 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.299612 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.299654 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.347827 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.571849 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vmlsp"] Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.910385 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.910889 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:06:40 crc kubenswrapper[4735]: I1122 08:06:40.949729 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:06:41 crc kubenswrapper[4735]: I1122 08:06:41.080289 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vmlsp" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerName="registry-server" containerID="cri-o://514893f8c26757a4d29715d50854383eed3b183826d3c44ceeccda96219694b9" gracePeriod=2 Nov 22 08:06:41 crc kubenswrapper[4735]: I1122 08:06:41.080711 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zmzcc" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerName="registry-server" containerID="cri-o://9aeada7369d1f7a19d4885baec8c1ce94eca5fa3f8f397501ae5592137fa640d" gracePeriod=2 Nov 22 08:06:41 crc kubenswrapper[4735]: I1122 08:06:41.116891 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:06:41 crc kubenswrapper[4735]: I1122 08:06:41.118077 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:06:41 crc kubenswrapper[4735]: I1122 08:06:41.118178 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:06:41 crc kubenswrapper[4735]: I1122 08:06:41.290083 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:06:41 crc kubenswrapper[4735]: I1122 08:06:41.290129 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:06:41 crc kubenswrapper[4735]: I1122 08:06:41.331401 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.086544 4735 generic.go:334] "Generic (PLEG): container finished" podID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerID="514893f8c26757a4d29715d50854383eed3b183826d3c44ceeccda96219694b9" exitCode=0 Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.086643 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vmlsp" event={"ID":"c97e92b6-f3cc-4be0-8931-5856d3d83097","Type":"ContainerDied","Data":"514893f8c26757a4d29715d50854383eed3b183826d3c44ceeccda96219694b9"} Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.093858 4735 generic.go:334] "Generic (PLEG): container finished" podID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerID="9aeada7369d1f7a19d4885baec8c1ce94eca5fa3f8f397501ae5592137fa640d" exitCode=0 Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.093887 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmzcc" event={"ID":"dd997979-7267-47a2-b486-1b34b36ac6a9","Type":"ContainerDied","Data":"9aeada7369d1f7a19d4885baec8c1ce94eca5fa3f8f397501ae5592137fa640d"} Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.153378 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.613972 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.620292 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.655518 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-utilities\") pod \"c97e92b6-f3cc-4be0-8931-5856d3d83097\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.655690 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-utilities\") pod \"dd997979-7267-47a2-b486-1b34b36ac6a9\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.655713 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-catalog-content\") pod \"dd997979-7267-47a2-b486-1b34b36ac6a9\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.655946 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-catalog-content\") pod \"c97e92b6-f3cc-4be0-8931-5856d3d83097\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.656308 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-utilities" (OuterVolumeSpecName: "utilities") pod "c97e92b6-f3cc-4be0-8931-5856d3d83097" (UID: "c97e92b6-f3cc-4be0-8931-5856d3d83097"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.657745 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d6nc\" (UniqueName: \"kubernetes.io/projected/dd997979-7267-47a2-b486-1b34b36ac6a9-kube-api-access-5d6nc\") pod \"dd997979-7267-47a2-b486-1b34b36ac6a9\" (UID: \"dd997979-7267-47a2-b486-1b34b36ac6a9\") " Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.657774 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzcrb\" (UniqueName: \"kubernetes.io/projected/c97e92b6-f3cc-4be0-8931-5856d3d83097-kube-api-access-wzcrb\") pod \"c97e92b6-f3cc-4be0-8931-5856d3d83097\" (UID: \"c97e92b6-f3cc-4be0-8931-5856d3d83097\") " Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.658247 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.656556 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-utilities" (OuterVolumeSpecName: "utilities") pod "dd997979-7267-47a2-b486-1b34b36ac6a9" (UID: "dd997979-7267-47a2-b486-1b34b36ac6a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.663600 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c97e92b6-f3cc-4be0-8931-5856d3d83097-kube-api-access-wzcrb" (OuterVolumeSpecName: "kube-api-access-wzcrb") pod "c97e92b6-f3cc-4be0-8931-5856d3d83097" (UID: "c97e92b6-f3cc-4be0-8931-5856d3d83097"). InnerVolumeSpecName "kube-api-access-wzcrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.663833 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd997979-7267-47a2-b486-1b34b36ac6a9-kube-api-access-5d6nc" (OuterVolumeSpecName: "kube-api-access-5d6nc") pod "dd997979-7267-47a2-b486-1b34b36ac6a9" (UID: "dd997979-7267-47a2-b486-1b34b36ac6a9"). InnerVolumeSpecName "kube-api-access-5d6nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.725838 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c97e92b6-f3cc-4be0-8931-5856d3d83097" (UID: "c97e92b6-f3cc-4be0-8931-5856d3d83097"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.737002 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd997979-7267-47a2-b486-1b34b36ac6a9" (UID: "dd997979-7267-47a2-b486-1b34b36ac6a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.764659 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.764704 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd997979-7267-47a2-b486-1b34b36ac6a9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.764719 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97e92b6-f3cc-4be0-8931-5856d3d83097-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.764731 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d6nc\" (UniqueName: \"kubernetes.io/projected/dd997979-7267-47a2-b486-1b34b36ac6a9-kube-api-access-5d6nc\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.764744 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzcrb\" (UniqueName: \"kubernetes.io/projected/c97e92b6-f3cc-4be0-8931-5856d3d83097-kube-api-access-wzcrb\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:42 crc kubenswrapper[4735]: I1122 08:06:42.977593 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g58rz"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.102882 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vmlsp" event={"ID":"c97e92b6-f3cc-4be0-8931-5856d3d83097","Type":"ContainerDied","Data":"2eb4af751f31cdfddb88cc1868a443cbe7a7bfe07ad900e527c5ec1cd7917a47"} Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.102937 4735 scope.go:117] "RemoveContainer" containerID="514893f8c26757a4d29715d50854383eed3b183826d3c44ceeccda96219694b9" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.103041 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vmlsp" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.110190 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmzcc" event={"ID":"dd997979-7267-47a2-b486-1b34b36ac6a9","Type":"ContainerDied","Data":"e8dad67d03850cf2eb9d1f3533618fdf9eedcd7af9c84ba10c4b222edec108a9"} Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.111069 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmzcc" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.111232 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g58rz" podUID="83793793-6201-43a7-85da-9da18c70441f" containerName="registry-server" containerID="cri-o://8e5fae32fb0b7b9348272aacd905bc66b744fbd847d44795a7f89a638431813a" gracePeriod=2 Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.136046 4735 scope.go:117] "RemoveContainer" containerID="a79925bc3ff4d047d4d754580e520d760a920c6f26ad09f9298c9950a5ec636e" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.140658 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vmlsp"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.150628 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vmlsp"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.156714 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zmzcc"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.161919 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zmzcc"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.168608 4735 scope.go:117] "RemoveContainer" containerID="734efa8c79d12cff03bfde2bd3640f0c0b30308c2737882309ec3158c04c9e83" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.188712 4735 scope.go:117] "RemoveContainer" containerID="9aeada7369d1f7a19d4885baec8c1ce94eca5fa3f8f397501ae5592137fa640d" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.209735 4735 scope.go:117] "RemoveContainer" containerID="ad459ac5d0f4dc394efbb2624eaf4ba9037b13470b81c82f39b3f11451605aca" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.232809 4735 scope.go:117] "RemoveContainer" containerID="a632959eb1c747edbf0c2e0f0f86cbcb622853ca3ee726dbd9b151be0868623d" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.271106 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" path="/var/lib/kubelet/pods/c97e92b6-f3cc-4be0-8931-5856d3d83097/volumes" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.271851 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" path="/var/lib/kubelet/pods/dd997979-7267-47a2-b486-1b34b36ac6a9/volumes" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.832619 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pc9bj"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.864944 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24jg6"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.865182 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-24jg6" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="registry-server" containerID="cri-o://ff584f951c30542bb3daaaa779121aa167a93c9a8ae1580349955fcdd31e9105" gracePeriod=30 Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.872832 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fqrjq"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.873063 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fqrjq" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerName="registry-server" containerID="cri-o://b2ce3e5ede47208f510fd30065bd8cb855d944474eefe9b12794d378a6cf9bb5" gracePeriod=30 Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.893246 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7k7n"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.893715 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" podUID="beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" containerName="marketplace-operator" containerID="cri-o://20d3b007d12f2d1240dfdc25605e42412fe3b015a6507a07d52bf7ce9317d5f1" gracePeriod=30 Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.905601 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl92s"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.905867 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bl92s" podUID="331e493d-6996-42ea-84ce-d388ab643352" containerName="registry-server" containerID="cri-o://111109e83efbee8881aa414acc422c0bb3fd638f020b5ad28a80491e10da537d" gracePeriod=30 Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.924404 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d6s67"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.924683 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d6s67" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerName="registry-server" containerID="cri-o://f33c76a4ae3a9f1004dba5a492066597b1df427f03d1c7741ed2e1091b44b611" gracePeriod=30 Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.933228 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f7s4z"] Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977248 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qr6cq"] Nov 22 08:06:43 crc kubenswrapper[4735]: E1122 08:06:43.977583 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerName="registry-server" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977595 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerName="registry-server" Nov 22 08:06:43 crc kubenswrapper[4735]: E1122 08:06:43.977611 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerName="extract-utilities" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977619 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerName="extract-utilities" Nov 22 08:06:43 crc kubenswrapper[4735]: E1122 08:06:43.977633 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerName="extract-content" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977641 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerName="extract-content" Nov 22 08:06:43 crc kubenswrapper[4735]: E1122 08:06:43.977655 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerName="registry-server" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977661 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerName="registry-server" Nov 22 08:06:43 crc kubenswrapper[4735]: E1122 08:06:43.977675 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerName="extract-utilities" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977681 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerName="extract-utilities" Nov 22 08:06:43 crc kubenswrapper[4735]: E1122 08:06:43.977691 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerName="extract-content" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977697 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerName="extract-content" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977857 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c97e92b6-f3cc-4be0-8931-5856d3d83097" containerName="registry-server" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.977873 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd997979-7267-47a2-b486-1b34b36ac6a9" containerName="registry-server" Nov 22 08:06:43 crc kubenswrapper[4735]: I1122 08:06:43.978363 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.053718 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qr6cq"] Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.089142 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrppt\" (UniqueName: \"kubernetes.io/projected/f10ff5dd-2bc5-4b07-90d2-174b896191bd-kube-api-access-jrppt\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.089479 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f10ff5dd-2bc5-4b07-90d2-174b896191bd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.089710 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f10ff5dd-2bc5-4b07-90d2-174b896191bd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.116818 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f7s4z" podUID="0e928195-0960-4cef-996e-35cc633b7a86" containerName="registry-server" containerID="cri-o://1092958533a50e85699c2f320db970bf10c84f45745a11fa07fe4fb65a4cf586" gracePeriod=30 Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.190623 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f10ff5dd-2bc5-4b07-90d2-174b896191bd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.190721 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f10ff5dd-2bc5-4b07-90d2-174b896191bd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.190750 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrppt\" (UniqueName: \"kubernetes.io/projected/f10ff5dd-2bc5-4b07-90d2-174b896191bd-kube-api-access-jrppt\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.191877 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f10ff5dd-2bc5-4b07-90d2-174b896191bd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.196111 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f10ff5dd-2bc5-4b07-90d2-174b896191bd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.221870 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrppt\" (UniqueName: \"kubernetes.io/projected/f10ff5dd-2bc5-4b07-90d2-174b896191bd-kube-api-access-jrppt\") pod \"marketplace-operator-79b997595-qr6cq\" (UID: \"f10ff5dd-2bc5-4b07-90d2-174b896191bd\") " pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.305686 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:44 crc kubenswrapper[4735]: I1122 08:06:44.540473 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qr6cq"] Nov 22 08:06:44 crc kubenswrapper[4735]: W1122 08:06:44.545491 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf10ff5dd_2bc5_4b07_90d2_174b896191bd.slice/crio-9bd45841bdabb958d94767d1f7d3e180aad635c57228ea35df2190deef3d5ec1 WatchSource:0}: Error finding container 9bd45841bdabb958d94767d1f7d3e180aad635c57228ea35df2190deef3d5ec1: Status 404 returned error can't find the container with id 9bd45841bdabb958d94767d1f7d3e180aad635c57228ea35df2190deef3d5ec1 Nov 22 08:06:45 crc kubenswrapper[4735]: I1122 08:06:45.122962 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" event={"ID":"f10ff5dd-2bc5-4b07-90d2-174b896191bd","Type":"ContainerStarted","Data":"9bd45841bdabb958d94767d1f7d3e180aad635c57228ea35df2190deef3d5ec1"} Nov 22 08:06:45 crc kubenswrapper[4735]: I1122 08:06:45.373582 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f7s4z"] Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.132084 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.132156 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.132210 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.133285 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.133448 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213" gracePeriod=600 Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.134133 4735 generic.go:334] "Generic (PLEG): container finished" podID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerID="b2ce3e5ede47208f510fd30065bd8cb855d944474eefe9b12794d378a6cf9bb5" exitCode=0 Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.134249 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqrjq" event={"ID":"3e5dab20-71fe-425e-b8e0-972f1ac0149a","Type":"ContainerDied","Data":"b2ce3e5ede47208f510fd30065bd8cb855d944474eefe9b12794d378a6cf9bb5"} Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.139997 4735 generic.go:334] "Generic (PLEG): container finished" podID="beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" containerID="20d3b007d12f2d1240dfdc25605e42412fe3b015a6507a07d52bf7ce9317d5f1" exitCode=0 Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.140133 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" event={"ID":"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3","Type":"ContainerDied","Data":"20d3b007d12f2d1240dfdc25605e42412fe3b015a6507a07d52bf7ce9317d5f1"} Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.151220 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" event={"ID":"f10ff5dd-2bc5-4b07-90d2-174b896191bd","Type":"ContainerStarted","Data":"84f4f0ef1e68266539f4f60053bf49f5032045e7710d4d81a3486950cab4cdb3"} Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.151531 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.153648 4735 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qr6cq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.153705 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" podUID="f10ff5dd-2bc5-4b07-90d2-174b896191bd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.159685 4735 generic.go:334] "Generic (PLEG): container finished" podID="331e493d-6996-42ea-84ce-d388ab643352" containerID="111109e83efbee8881aa414acc422c0bb3fd638f020b5ad28a80491e10da537d" exitCode=0 Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.159742 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl92s" event={"ID":"331e493d-6996-42ea-84ce-d388ab643352","Type":"ContainerDied","Data":"111109e83efbee8881aa414acc422c0bb3fd638f020b5ad28a80491e10da537d"} Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.186555 4735 generic.go:334] "Generic (PLEG): container finished" podID="cdab44ae-77cb-454d-b040-c677adfda315" containerID="ff584f951c30542bb3daaaa779121aa167a93c9a8ae1580349955fcdd31e9105" exitCode=0 Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.186632 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24jg6" event={"ID":"cdab44ae-77cb-454d-b040-c677adfda315","Type":"ContainerDied","Data":"ff584f951c30542bb3daaaa779121aa167a93c9a8ae1580349955fcdd31e9105"} Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.197847 4735 generic.go:334] "Generic (PLEG): container finished" podID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerID="f33c76a4ae3a9f1004dba5a492066597b1df427f03d1c7741ed2e1091b44b611" exitCode=0 Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.197917 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6s67" event={"ID":"f56f24fc-45bb-4842-9de9-2e0f955b04b0","Type":"ContainerDied","Data":"f33c76a4ae3a9f1004dba5a492066597b1df427f03d1c7741ed2e1091b44b611"} Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.240247 4735 generic.go:334] "Generic (PLEG): container finished" podID="0e928195-0960-4cef-996e-35cc633b7a86" containerID="1092958533a50e85699c2f320db970bf10c84f45745a11fa07fe4fb65a4cf586" exitCode=0 Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.240340 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f7s4z" event={"ID":"0e928195-0960-4cef-996e-35cc633b7a86","Type":"ContainerDied","Data":"1092958533a50e85699c2f320db970bf10c84f45745a11fa07fe4fb65a4cf586"} Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.247301 4735 generic.go:334] "Generic (PLEG): container finished" podID="83793793-6201-43a7-85da-9da18c70441f" containerID="8e5fae32fb0b7b9348272aacd905bc66b744fbd847d44795a7f89a638431813a" exitCode=0 Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.247351 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g58rz" event={"ID":"83793793-6201-43a7-85da-9da18c70441f","Type":"ContainerDied","Data":"8e5fae32fb0b7b9348272aacd905bc66b744fbd847d44795a7f89a638431813a"} Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.367212 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.396364 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" podStartSLOduration=3.396339598 podStartE2EDuration="3.396339598s" podCreationTimestamp="2025-11-22 08:06:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:06:46.174937126 +0000 UTC m=+227.779275731" watchObservedRunningTime="2025-11-22 08:06:46.396339598 +0000 UTC m=+228.000678203" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.460454 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.467224 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-utilities\") pod \"cdab44ae-77cb-454d-b040-c677adfda315\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.467272 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-catalog-content\") pod \"cdab44ae-77cb-454d-b040-c677adfda315\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.467303 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzk8g\" (UniqueName: \"kubernetes.io/projected/cdab44ae-77cb-454d-b040-c677adfda315-kube-api-access-pzk8g\") pod \"cdab44ae-77cb-454d-b040-c677adfda315\" (UID: \"cdab44ae-77cb-454d-b040-c677adfda315\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.468062 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-utilities" (OuterVolumeSpecName: "utilities") pod "cdab44ae-77cb-454d-b040-c677adfda315" (UID: "cdab44ae-77cb-454d-b040-c677adfda315"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.469153 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.472807 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdab44ae-77cb-454d-b040-c677adfda315-kube-api-access-pzk8g" (OuterVolumeSpecName: "kube-api-access-pzk8g") pod "cdab44ae-77cb-454d-b040-c677adfda315" (UID: "cdab44ae-77cb-454d-b040-c677adfda315"). InnerVolumeSpecName "kube-api-access-pzk8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.549356 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cdab44ae-77cb-454d-b040-c677adfda315" (UID: "cdab44ae-77cb-454d-b040-c677adfda315"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.556121 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.568217 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-catalog-content\") pod \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.568291 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-utilities\") pod \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.568357 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-operator-metrics\") pod \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.568559 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-trusted-ca\") pod \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.568639 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb5ct\" (UniqueName: \"kubernetes.io/projected/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-kube-api-access-fb5ct\") pod \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\" (UID: \"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.568674 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtp4g\" (UniqueName: \"kubernetes.io/projected/3e5dab20-71fe-425e-b8e0-972f1ac0149a-kube-api-access-xtp4g\") pod \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\" (UID: \"3e5dab20-71fe-425e-b8e0-972f1ac0149a\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.570290 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.570338 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdab44ae-77cb-454d-b040-c677adfda315-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.570355 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzk8g\" (UniqueName: \"kubernetes.io/projected/cdab44ae-77cb-454d-b040-c677adfda315-kube-api-access-pzk8g\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.570363 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-utilities" (OuterVolumeSpecName: "utilities") pod "3e5dab20-71fe-425e-b8e0-972f1ac0149a" (UID: "3e5dab20-71fe-425e-b8e0-972f1ac0149a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.571097 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" (UID: "beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.573339 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e5dab20-71fe-425e-b8e0-972f1ac0149a-kube-api-access-xtp4g" (OuterVolumeSpecName: "kube-api-access-xtp4g") pod "3e5dab20-71fe-425e-b8e0-972f1ac0149a" (UID: "3e5dab20-71fe-425e-b8e0-972f1ac0149a"). InnerVolumeSpecName "kube-api-access-xtp4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.580917 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" (UID: "beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.581140 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-kube-api-access-fb5ct" (OuterVolumeSpecName: "kube-api-access-fb5ct") pod "beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" (UID: "beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3"). InnerVolumeSpecName "kube-api-access-fb5ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.653573 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e5dab20-71fe-425e-b8e0-972f1ac0149a" (UID: "3e5dab20-71fe-425e-b8e0-972f1ac0149a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671120 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-catalog-content\") pod \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671186 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-utilities\") pod \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671226 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcqqf\" (UniqueName: \"kubernetes.io/projected/f56f24fc-45bb-4842-9de9-2e0f955b04b0-kube-api-access-wcqqf\") pod \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\" (UID: \"f56f24fc-45bb-4842-9de9-2e0f955b04b0\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671511 4735 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671524 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb5ct\" (UniqueName: \"kubernetes.io/projected/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-kube-api-access-fb5ct\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671533 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtp4g\" (UniqueName: \"kubernetes.io/projected/3e5dab20-71fe-425e-b8e0-972f1ac0149a-kube-api-access-xtp4g\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671542 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671552 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5dab20-71fe-425e-b8e0-972f1ac0149a-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.671560 4735 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.672599 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-utilities" (OuterVolumeSpecName: "utilities") pod "f56f24fc-45bb-4842-9de9-2e0f955b04b0" (UID: "f56f24fc-45bb-4842-9de9-2e0f955b04b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.674095 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56f24fc-45bb-4842-9de9-2e0f955b04b0-kube-api-access-wcqqf" (OuterVolumeSpecName: "kube-api-access-wcqqf") pod "f56f24fc-45bb-4842-9de9-2e0f955b04b0" (UID: "f56f24fc-45bb-4842-9de9-2e0f955b04b0"). InnerVolumeSpecName "kube-api-access-wcqqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.702084 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.758973 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f56f24fc-45bb-4842-9de9-2e0f955b04b0" (UID: "f56f24fc-45bb-4842-9de9-2e0f955b04b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.772796 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.772830 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcqqf\" (UniqueName: \"kubernetes.io/projected/f56f24fc-45bb-4842-9de9-2e0f955b04b0-kube-api-access-wcqqf\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.772845 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56f24fc-45bb-4842-9de9-2e0f955b04b0-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.874244 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-catalog-content\") pod \"0e928195-0960-4cef-996e-35cc633b7a86\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.874334 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-utilities\") pod \"0e928195-0960-4cef-996e-35cc633b7a86\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.874356 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t64mq\" (UniqueName: \"kubernetes.io/projected/0e928195-0960-4cef-996e-35cc633b7a86-kube-api-access-t64mq\") pod \"0e928195-0960-4cef-996e-35cc633b7a86\" (UID: \"0e928195-0960-4cef-996e-35cc633b7a86\") " Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.877650 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e928195-0960-4cef-996e-35cc633b7a86-kube-api-access-t64mq" (OuterVolumeSpecName: "kube-api-access-t64mq") pod "0e928195-0960-4cef-996e-35cc633b7a86" (UID: "0e928195-0960-4cef-996e-35cc633b7a86"). InnerVolumeSpecName "kube-api-access-t64mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.881322 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-utilities" (OuterVolumeSpecName: "utilities") pod "0e928195-0960-4cef-996e-35cc633b7a86" (UID: "0e928195-0960-4cef-996e-35cc633b7a86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.975422 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.975731 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t64mq\" (UniqueName: \"kubernetes.io/projected/0e928195-0960-4cef-996e-35cc633b7a86-kube-api-access-t64mq\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:46 crc kubenswrapper[4735]: I1122 08:06:46.989508 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e928195-0960-4cef-996e-35cc633b7a86" (UID: "0e928195-0960-4cef-996e-35cc633b7a86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.011920 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.078478 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e928195-0960-4cef-996e-35cc633b7a86-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.088566 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.179130 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-catalog-content\") pod \"83793793-6201-43a7-85da-9da18c70441f\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.179215 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-utilities\") pod \"83793793-6201-43a7-85da-9da18c70441f\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.179253 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h96pl\" (UniqueName: \"kubernetes.io/projected/83793793-6201-43a7-85da-9da18c70441f-kube-api-access-h96pl\") pod \"83793793-6201-43a7-85da-9da18c70441f\" (UID: \"83793793-6201-43a7-85da-9da18c70441f\") " Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.179306 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2b4f\" (UniqueName: \"kubernetes.io/projected/331e493d-6996-42ea-84ce-d388ab643352-kube-api-access-k2b4f\") pod \"331e493d-6996-42ea-84ce-d388ab643352\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.179350 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-catalog-content\") pod \"331e493d-6996-42ea-84ce-d388ab643352\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.179376 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-utilities\") pod \"331e493d-6996-42ea-84ce-d388ab643352\" (UID: \"331e493d-6996-42ea-84ce-d388ab643352\") " Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.180120 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-utilities" (OuterVolumeSpecName: "utilities") pod "83793793-6201-43a7-85da-9da18c70441f" (UID: "83793793-6201-43a7-85da-9da18c70441f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.180375 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-utilities" (OuterVolumeSpecName: "utilities") pod "331e493d-6996-42ea-84ce-d388ab643352" (UID: "331e493d-6996-42ea-84ce-d388ab643352"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.183418 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83793793-6201-43a7-85da-9da18c70441f-kube-api-access-h96pl" (OuterVolumeSpecName: "kube-api-access-h96pl") pod "83793793-6201-43a7-85da-9da18c70441f" (UID: "83793793-6201-43a7-85da-9da18c70441f"). InnerVolumeSpecName "kube-api-access-h96pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.183503 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/331e493d-6996-42ea-84ce-d388ab643352-kube-api-access-k2b4f" (OuterVolumeSpecName: "kube-api-access-k2b4f") pod "331e493d-6996-42ea-84ce-d388ab643352" (UID: "331e493d-6996-42ea-84ce-d388ab643352"). InnerVolumeSpecName "kube-api-access-k2b4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.198018 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "331e493d-6996-42ea-84ce-d388ab643352" (UID: "331e493d-6996-42ea-84ce-d388ab643352"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.203188 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83793793-6201-43a7-85da-9da18c70441f" (UID: "83793793-6201-43a7-85da-9da18c70441f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.254850 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f7s4z" event={"ID":"0e928195-0960-4cef-996e-35cc633b7a86","Type":"ContainerDied","Data":"7c6b0e5956f758ce83691e8eadeed02f2f62b62131615a36ac2602f3179ff4cc"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.255150 4735 scope.go:117] "RemoveContainer" containerID="1092958533a50e85699c2f320db970bf10c84f45745a11fa07fe4fb65a4cf586" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.254932 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f7s4z" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.261009 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6s67" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.261085 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6s67" event={"ID":"f56f24fc-45bb-4842-9de9-2e0f955b04b0","Type":"ContainerDied","Data":"de73a2ad05299894354afdad0d995eb3cbe6be667e3bceae589f430575078f1c"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.266258 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g58rz" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.268999 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqrjq" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.272406 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.277581 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g58rz" event={"ID":"83793793-6201-43a7-85da-9da18c70441f","Type":"ContainerDied","Data":"497e578526be7e1fd075fb274db306777fdf34518037e26de118b67ac23e8d82"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.277624 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqrjq" event={"ID":"3e5dab20-71fe-425e-b8e0-972f1ac0149a","Type":"ContainerDied","Data":"c6826058da78bb3e0e6020230d4323151edebc2ce5a2bf849f477f23de3face8"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.277642 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v7k7n" event={"ID":"beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3","Type":"ContainerDied","Data":"515965c2af5d1e9d156c0ae801d8fd7d2a600fe4070d12da588fa722595094d0"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.280821 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213" exitCode=0 Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.280961 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.281024 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"c483bef07756ce17b03463d41a2cb38d389b0b284cac0a615aac582d5c89aec3"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.285980 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.286020 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331e493d-6996-42ea-84ce-d388ab643352-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.286033 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.286044 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83793793-6201-43a7-85da-9da18c70441f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.286058 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h96pl\" (UniqueName: \"kubernetes.io/projected/83793793-6201-43a7-85da-9da18c70441f-kube-api-access-h96pl\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.286072 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2b4f\" (UniqueName: \"kubernetes.io/projected/331e493d-6996-42ea-84ce-d388ab643352-kube-api-access-k2b4f\") on node \"crc\" DevicePath \"\"" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.291859 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl92s" event={"ID":"331e493d-6996-42ea-84ce-d388ab643352","Type":"ContainerDied","Data":"9f2f213b46e9dd2563085f510c92c20b7827d1f83674e08ac441e4c73912e510"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.291891 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bl92s" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.295564 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24jg6" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.302218 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24jg6" event={"ID":"cdab44ae-77cb-454d-b040-c677adfda315","Type":"ContainerDied","Data":"b27b394e3d579a6d8393b154549e20e7425d84e366001332b7d2754c582ec867"} Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.304750 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qr6cq" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.315013 4735 scope.go:117] "RemoveContainer" containerID="c1337b61a02682a1ac56eb2d5236e6b563d07aec58608852aa4364e23c692637" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.351170 4735 scope.go:117] "RemoveContainer" containerID="49ea8b71b387b9646b41aedda6b3a72a9eed818abc024568302e89d919c63aa4" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.397680 4735 scope.go:117] "RemoveContainer" containerID="f33c76a4ae3a9f1004dba5a492066597b1df427f03d1c7741ed2e1091b44b611" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.403023 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f7s4z"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.409557 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f7s4z"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.413880 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl92s"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.420842 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl92s"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.421447 4735 scope.go:117] "RemoveContainer" containerID="aaee914902b1945e64e472ac5b0606fb6291dc61c1fe4aaba33176db07cee75d" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.424170 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d6s67"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.430497 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d6s67"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.440901 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g58rz"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.440951 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g58rz"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.455573 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24jg6"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.455858 4735 scope.go:117] "RemoveContainer" containerID="033bee9303354f6215d8e67cb5549e7322d871a760cf20faab81bcf715fa2b7c" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.457761 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-24jg6"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.461292 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fqrjq"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.464495 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fqrjq"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.477022 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7k7n"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.477588 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7k7n"] Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.479536 4735 scope.go:117] "RemoveContainer" containerID="8e5fae32fb0b7b9348272aacd905bc66b744fbd847d44795a7f89a638431813a" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.491527 4735 scope.go:117] "RemoveContainer" containerID="5aa828dd9baa214f29a92743a91485338e102ff2d6267349dabe165da7ef3f0e" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.504094 4735 scope.go:117] "RemoveContainer" containerID="10bd26c2744f3e317ab99bf7d36620e74923b6461ca998371cc3885ce6d169df" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.514414 4735 scope.go:117] "RemoveContainer" containerID="b2ce3e5ede47208f510fd30065bd8cb855d944474eefe9b12794d378a6cf9bb5" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.525686 4735 scope.go:117] "RemoveContainer" containerID="6eae83fae1169b46961a643ee4bf4d38a5e5766dad95d8669eae379262e59b45" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.540222 4735 scope.go:117] "RemoveContainer" containerID="7cbff1501ca554f917bd325be6e3d45200555f8979ccf2b95055f38c7a3f092d" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.559322 4735 scope.go:117] "RemoveContainer" containerID="20d3b007d12f2d1240dfdc25605e42412fe3b015a6507a07d52bf7ce9317d5f1" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.581008 4735 scope.go:117] "RemoveContainer" containerID="111109e83efbee8881aa414acc422c0bb3fd638f020b5ad28a80491e10da537d" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.593838 4735 scope.go:117] "RemoveContainer" containerID="16873a97676e311f9b97c000d1dd788b7dd996b293d1de9da5a638a9be08f9ed" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.615623 4735 scope.go:117] "RemoveContainer" containerID="5db4815b791297c8c59efb01d3bccbdf2306095f0b249a5f43044ffbf9fb37ed" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.631144 4735 scope.go:117] "RemoveContainer" containerID="ff584f951c30542bb3daaaa779121aa167a93c9a8ae1580349955fcdd31e9105" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.646158 4735 scope.go:117] "RemoveContainer" containerID="b283e7e776d674d16413252df4861bbc2860eee2c68ba0d765a7b3959632fb77" Nov 22 08:06:47 crc kubenswrapper[4735]: I1122 08:06:47.665218 4735 scope.go:117] "RemoveContainer" containerID="993efa0fa6d9a52b6e2cc3641d0534abbabba3c41442443e727a333bd6acadff" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.822399 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l4fxx"] Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.822931 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.822946 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.822960 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.822970 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.822981 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83793793-6201-43a7-85da-9da18c70441f" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.822990 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="83793793-6201-43a7-85da-9da18c70441f" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.822999 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823006 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823017 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331e493d-6996-42ea-84ce-d388ab643352" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823024 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="331e493d-6996-42ea-84ce-d388ab643352" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823035 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331e493d-6996-42ea-84ce-d388ab643352" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823044 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="331e493d-6996-42ea-84ce-d388ab643352" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823053 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83793793-6201-43a7-85da-9da18c70441f" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823061 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="83793793-6201-43a7-85da-9da18c70441f" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823071 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823079 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823090 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e928195-0960-4cef-996e-35cc633b7a86" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823097 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e928195-0960-4cef-996e-35cc633b7a86" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823111 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823119 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823126 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823133 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823146 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823153 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823164 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e928195-0960-4cef-996e-35cc633b7a86" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823171 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e928195-0960-4cef-996e-35cc633b7a86" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823179 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e928195-0960-4cef-996e-35cc633b7a86" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823187 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e928195-0960-4cef-996e-35cc633b7a86" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823196 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823203 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823215 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83793793-6201-43a7-85da-9da18c70441f" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823223 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="83793793-6201-43a7-85da-9da18c70441f" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823233 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823241 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerName="extract-content" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823251 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331e493d-6996-42ea-84ce-d388ab643352" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823259 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="331e493d-6996-42ea-84ce-d388ab643352" containerName="extract-utilities" Nov 22 08:06:48 crc kubenswrapper[4735]: E1122 08:06:48.823270 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" containerName="marketplace-operator" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823278 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" containerName="marketplace-operator" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823384 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823401 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e928195-0960-4cef-996e-35cc633b7a86" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823414 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823426 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="83793793-6201-43a7-85da-9da18c70441f" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823435 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="331e493d-6996-42ea-84ce-d388ab643352" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823445 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" containerName="marketplace-operator" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.823474 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdab44ae-77cb-454d-b040-c677adfda315" containerName="registry-server" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.824415 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.827220 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.837269 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l4fxx"] Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.904896 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6dab36d-6121-4c4b-b350-7594bc1c0133-utilities\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.904960 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6dab36d-6121-4c4b-b350-7594bc1c0133-catalog-content\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:48 crc kubenswrapper[4735]: I1122 08:06:48.905014 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9wrh\" (UniqueName: \"kubernetes.io/projected/c6dab36d-6121-4c4b-b350-7594bc1c0133-kube-api-access-h9wrh\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.005656 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9wrh\" (UniqueName: \"kubernetes.io/projected/c6dab36d-6121-4c4b-b350-7594bc1c0133-kube-api-access-h9wrh\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.005723 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6dab36d-6121-4c4b-b350-7594bc1c0133-utilities\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.005761 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6dab36d-6121-4c4b-b350-7594bc1c0133-catalog-content\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.006274 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6dab36d-6121-4c4b-b350-7594bc1c0133-catalog-content\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.006345 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6dab36d-6121-4c4b-b350-7594bc1c0133-utilities\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.025903 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9wrh\" (UniqueName: \"kubernetes.io/projected/c6dab36d-6121-4c4b-b350-7594bc1c0133-kube-api-access-h9wrh\") pod \"certified-operators-l4fxx\" (UID: \"c6dab36d-6121-4c4b-b350-7594bc1c0133\") " pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.161261 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.275575 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e928195-0960-4cef-996e-35cc633b7a86" path="/var/lib/kubelet/pods/0e928195-0960-4cef-996e-35cc633b7a86/volumes" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.277382 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="331e493d-6996-42ea-84ce-d388ab643352" path="/var/lib/kubelet/pods/331e493d-6996-42ea-84ce-d388ab643352/volumes" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.278774 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e5dab20-71fe-425e-b8e0-972f1ac0149a" path="/var/lib/kubelet/pods/3e5dab20-71fe-425e-b8e0-972f1ac0149a/volumes" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.280981 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83793793-6201-43a7-85da-9da18c70441f" path="/var/lib/kubelet/pods/83793793-6201-43a7-85da-9da18c70441f/volumes" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.282850 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3" path="/var/lib/kubelet/pods/beb4e0bc-1b08-491c-bcf2-cbd1bc2507f3/volumes" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.284750 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdab44ae-77cb-454d-b040-c677adfda315" path="/var/lib/kubelet/pods/cdab44ae-77cb-454d-b040-c677adfda315/volumes" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.286100 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56f24fc-45bb-4842-9de9-2e0f955b04b0" path="/var/lib/kubelet/pods/f56f24fc-45bb-4842-9de9-2e0f955b04b0/volumes" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.360508 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l4fxx"] Nov 22 08:06:49 crc kubenswrapper[4735]: W1122 08:06:49.371523 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6dab36d_6121_4c4b_b350_7594bc1c0133.slice/crio-04d2087a5a8f694c0e4dcb34d04f39be2e7776e84101324992a38486d3b75836 WatchSource:0}: Error finding container 04d2087a5a8f694c0e4dcb34d04f39be2e7776e84101324992a38486d3b75836: Status 404 returned error can't find the container with id 04d2087a5a8f694c0e4dcb34d04f39be2e7776e84101324992a38486d3b75836 Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.784963 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2cq5w"] Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.797177 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2cq5w"] Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.797306 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.800192 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.917591 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-utilities\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.917719 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djp4s\" (UniqueName: \"kubernetes.io/projected/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-kube-api-access-djp4s\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:49 crc kubenswrapper[4735]: I1122 08:06:49.917813 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-catalog-content\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.019354 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djp4s\" (UniqueName: \"kubernetes.io/projected/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-kube-api-access-djp4s\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.019514 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-catalog-content\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.019555 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-utilities\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.020317 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-catalog-content\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.020324 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-utilities\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.041235 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djp4s\" (UniqueName: \"kubernetes.io/projected/76c916a7-a1da-48e5-b7cd-1827b4d42c7a-kube-api-access-djp4s\") pod \"redhat-operators-2cq5w\" (UID: \"76c916a7-a1da-48e5-b7cd-1827b4d42c7a\") " pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.122972 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.304836 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2cq5w"] Nov 22 08:06:50 crc kubenswrapper[4735]: W1122 08:06:50.311099 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76c916a7_a1da_48e5_b7cd_1827b4d42c7a.slice/crio-b59d5e43ef6c8ec01aeec5022839985498a40902ca5f6bcf3a992e8265cf8934 WatchSource:0}: Error finding container b59d5e43ef6c8ec01aeec5022839985498a40902ca5f6bcf3a992e8265cf8934: Status 404 returned error can't find the container with id b59d5e43ef6c8ec01aeec5022839985498a40902ca5f6bcf3a992e8265cf8934 Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.347925 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cq5w" event={"ID":"76c916a7-a1da-48e5-b7cd-1827b4d42c7a","Type":"ContainerStarted","Data":"b59d5e43ef6c8ec01aeec5022839985498a40902ca5f6bcf3a992e8265cf8934"} Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.350131 4735 generic.go:334] "Generic (PLEG): container finished" podID="c6dab36d-6121-4c4b-b350-7594bc1c0133" containerID="cea2ae40f26dd60441d22ce9bb34fe221888aa47774cc32bdf35e0e0a7da79d4" exitCode=0 Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.350180 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4fxx" event={"ID":"c6dab36d-6121-4c4b-b350-7594bc1c0133","Type":"ContainerDied","Data":"cea2ae40f26dd60441d22ce9bb34fe221888aa47774cc32bdf35e0e0a7da79d4"} Nov 22 08:06:50 crc kubenswrapper[4735]: I1122 08:06:50.350228 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4fxx" event={"ID":"c6dab36d-6121-4c4b-b350-7594bc1c0133","Type":"ContainerStarted","Data":"04d2087a5a8f694c0e4dcb34d04f39be2e7776e84101324992a38486d3b75836"} Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.179255 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p8qf6"] Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.180827 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.183170 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.193012 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p8qf6"] Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.335976 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70cac5fe-7dac-4acc-87a0-c4194fdae63c-catalog-content\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.336024 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70cac5fe-7dac-4acc-87a0-c4194fdae63c-utilities\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.336121 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7gvt\" (UniqueName: \"kubernetes.io/projected/70cac5fe-7dac-4acc-87a0-c4194fdae63c-kube-api-access-t7gvt\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.357667 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4fxx" event={"ID":"c6dab36d-6121-4c4b-b350-7594bc1c0133","Type":"ContainerStarted","Data":"dfe3debd516d16f7bbdf8f11db131895afdc15a5391d11e06e2780874b483fcb"} Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.359530 4735 generic.go:334] "Generic (PLEG): container finished" podID="76c916a7-a1da-48e5-b7cd-1827b4d42c7a" containerID="28ce02a05170e0a623c2570cb6e26765c87686ceb0eba4fb0b373bd85216714e" exitCode=0 Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.359558 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cq5w" event={"ID":"76c916a7-a1da-48e5-b7cd-1827b4d42c7a","Type":"ContainerDied","Data":"28ce02a05170e0a623c2570cb6e26765c87686ceb0eba4fb0b373bd85216714e"} Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.437029 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70cac5fe-7dac-4acc-87a0-c4194fdae63c-catalog-content\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.437085 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70cac5fe-7dac-4acc-87a0-c4194fdae63c-utilities\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.437134 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7gvt\" (UniqueName: \"kubernetes.io/projected/70cac5fe-7dac-4acc-87a0-c4194fdae63c-kube-api-access-t7gvt\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.437952 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70cac5fe-7dac-4acc-87a0-c4194fdae63c-catalog-content\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.439362 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70cac5fe-7dac-4acc-87a0-c4194fdae63c-utilities\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.456698 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7gvt\" (UniqueName: \"kubernetes.io/projected/70cac5fe-7dac-4acc-87a0-c4194fdae63c-kube-api-access-t7gvt\") pod \"community-operators-p8qf6\" (UID: \"70cac5fe-7dac-4acc-87a0-c4194fdae63c\") " pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.499804 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:06:51 crc kubenswrapper[4735]: I1122 08:06:51.709268 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p8qf6"] Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.178038 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p7s"] Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.179269 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.185952 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.190400 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p7s"] Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.249798 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1964504-b509-4e0e-b88e-2577618a493a-catalog-content\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.249939 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qddsm\" (UniqueName: \"kubernetes.io/projected/f1964504-b509-4e0e-b88e-2577618a493a-kube-api-access-qddsm\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.250090 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1964504-b509-4e0e-b88e-2577618a493a-utilities\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.351815 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qddsm\" (UniqueName: \"kubernetes.io/projected/f1964504-b509-4e0e-b88e-2577618a493a-kube-api-access-qddsm\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.351888 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1964504-b509-4e0e-b88e-2577618a493a-utilities\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.352027 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1964504-b509-4e0e-b88e-2577618a493a-catalog-content\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.352703 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1964504-b509-4e0e-b88e-2577618a493a-utilities\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.353043 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1964504-b509-4e0e-b88e-2577618a493a-catalog-content\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.372657 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qddsm\" (UniqueName: \"kubernetes.io/projected/f1964504-b509-4e0e-b88e-2577618a493a-kube-api-access-qddsm\") pod \"redhat-marketplace-h5p7s\" (UID: \"f1964504-b509-4e0e-b88e-2577618a493a\") " pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.374237 4735 generic.go:334] "Generic (PLEG): container finished" podID="70cac5fe-7dac-4acc-87a0-c4194fdae63c" containerID="e4018f07bd5fcbe79b65e48d708f17daf05b55563a781cae3bc98331da1bf32a" exitCode=0 Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.374302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8qf6" event={"ID":"70cac5fe-7dac-4acc-87a0-c4194fdae63c","Type":"ContainerDied","Data":"e4018f07bd5fcbe79b65e48d708f17daf05b55563a781cae3bc98331da1bf32a"} Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.374335 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8qf6" event={"ID":"70cac5fe-7dac-4acc-87a0-c4194fdae63c","Type":"ContainerStarted","Data":"dc5006a2b46c76af5242fa565ff22e81c4453f4c8fed95c7c120e067f74b181f"} Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.378691 4735 generic.go:334] "Generic (PLEG): container finished" podID="c6dab36d-6121-4c4b-b350-7594bc1c0133" containerID="dfe3debd516d16f7bbdf8f11db131895afdc15a5391d11e06e2780874b483fcb" exitCode=0 Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.378732 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4fxx" event={"ID":"c6dab36d-6121-4c4b-b350-7594bc1c0133","Type":"ContainerDied","Data":"dfe3debd516d16f7bbdf8f11db131895afdc15a5391d11e06e2780874b483fcb"} Nov 22 08:06:52 crc kubenswrapper[4735]: I1122 08:06:52.511265 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:06:53 crc kubenswrapper[4735]: I1122 08:06:53.090298 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p7s"] Nov 22 08:06:53 crc kubenswrapper[4735]: W1122 08:06:53.091442 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1964504_b509_4e0e_b88e_2577618a493a.slice/crio-edf08829aca2ebfdd2d2df6987038310eea39cf527ca07930a2b91dfd96037f0 WatchSource:0}: Error finding container edf08829aca2ebfdd2d2df6987038310eea39cf527ca07930a2b91dfd96037f0: Status 404 returned error can't find the container with id edf08829aca2ebfdd2d2df6987038310eea39cf527ca07930a2b91dfd96037f0 Nov 22 08:06:53 crc kubenswrapper[4735]: I1122 08:06:53.389366 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4fxx" event={"ID":"c6dab36d-6121-4c4b-b350-7594bc1c0133","Type":"ContainerStarted","Data":"00604eaac8d0bf0d571a604e755c49b6cace4ff5eeab995feaf39e8cff761215"} Nov 22 08:06:53 crc kubenswrapper[4735]: I1122 08:06:53.393821 4735 generic.go:334] "Generic (PLEG): container finished" podID="f1964504-b509-4e0e-b88e-2577618a493a" containerID="e82e1a6a1a6eb609a740af6265f563a83c05f40abc2863ede84e52abfd9f9061" exitCode=0 Nov 22 08:06:53 crc kubenswrapper[4735]: I1122 08:06:53.393895 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p7s" event={"ID":"f1964504-b509-4e0e-b88e-2577618a493a","Type":"ContainerDied","Data":"e82e1a6a1a6eb609a740af6265f563a83c05f40abc2863ede84e52abfd9f9061"} Nov 22 08:06:53 crc kubenswrapper[4735]: I1122 08:06:53.393923 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p7s" event={"ID":"f1964504-b509-4e0e-b88e-2577618a493a","Type":"ContainerStarted","Data":"edf08829aca2ebfdd2d2df6987038310eea39cf527ca07930a2b91dfd96037f0"} Nov 22 08:06:53 crc kubenswrapper[4735]: I1122 08:06:53.397994 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cq5w" event={"ID":"76c916a7-a1da-48e5-b7cd-1827b4d42c7a","Type":"ContainerStarted","Data":"ddbf5f5d6588bec12414cdab92a121401aa4879de4a660309841edcf7d8d56ce"} Nov 22 08:06:53 crc kubenswrapper[4735]: I1122 08:06:53.414206 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l4fxx" podStartSLOduration=2.52285119 podStartE2EDuration="5.414182485s" podCreationTimestamp="2025-11-22 08:06:48 +0000 UTC" firstStartedPulling="2025-11-22 08:06:50.352274669 +0000 UTC m=+231.956613304" lastFinishedPulling="2025-11-22 08:06:53.243605994 +0000 UTC m=+234.847944599" observedRunningTime="2025-11-22 08:06:53.410499336 +0000 UTC m=+235.014837941" watchObservedRunningTime="2025-11-22 08:06:53.414182485 +0000 UTC m=+235.018521100" Nov 22 08:06:54 crc kubenswrapper[4735]: I1122 08:06:54.027002 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-wmm22" Nov 22 08:06:54 crc kubenswrapper[4735]: I1122 08:06:54.148794 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9r74q"] Nov 22 08:06:54 crc kubenswrapper[4735]: I1122 08:06:54.404876 4735 generic.go:334] "Generic (PLEG): container finished" podID="70cac5fe-7dac-4acc-87a0-c4194fdae63c" containerID="e9a5eab106af1f066795807c1450bcd08e122766e9f70c3c4fa5e8395b71b380" exitCode=0 Nov 22 08:06:54 crc kubenswrapper[4735]: I1122 08:06:54.404968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8qf6" event={"ID":"70cac5fe-7dac-4acc-87a0-c4194fdae63c","Type":"ContainerDied","Data":"e9a5eab106af1f066795807c1450bcd08e122766e9f70c3c4fa5e8395b71b380"} Nov 22 08:06:54 crc kubenswrapper[4735]: I1122 08:06:54.409861 4735 generic.go:334] "Generic (PLEG): container finished" podID="76c916a7-a1da-48e5-b7cd-1827b4d42c7a" containerID="ddbf5f5d6588bec12414cdab92a121401aa4879de4a660309841edcf7d8d56ce" exitCode=0 Nov 22 08:06:54 crc kubenswrapper[4735]: I1122 08:06:54.409928 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cq5w" event={"ID":"76c916a7-a1da-48e5-b7cd-1827b4d42c7a","Type":"ContainerDied","Data":"ddbf5f5d6588bec12414cdab92a121401aa4879de4a660309841edcf7d8d56ce"} Nov 22 08:06:54 crc kubenswrapper[4735]: I1122 08:06:54.413206 4735 generic.go:334] "Generic (PLEG): container finished" podID="f1964504-b509-4e0e-b88e-2577618a493a" containerID="6ac5d91a3e30d7ca4362cb6414a31b526449cb287924f99dc3fb6a1653e4a14b" exitCode=0 Nov 22 08:06:54 crc kubenswrapper[4735]: I1122 08:06:54.413255 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p7s" event={"ID":"f1964504-b509-4e0e-b88e-2577618a493a","Type":"ContainerDied","Data":"6ac5d91a3e30d7ca4362cb6414a31b526449cb287924f99dc3fb6a1653e4a14b"} Nov 22 08:06:55 crc kubenswrapper[4735]: I1122 08:06:55.419750 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p7s" event={"ID":"f1964504-b509-4e0e-b88e-2577618a493a","Type":"ContainerStarted","Data":"cb760ca00173beffa90c4043e6ac025d9c04c4e284dd3393b6def7aa33af3d06"} Nov 22 08:06:55 crc kubenswrapper[4735]: I1122 08:06:55.422499 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8qf6" event={"ID":"70cac5fe-7dac-4acc-87a0-c4194fdae63c","Type":"ContainerStarted","Data":"a3c8c35ccc7f6ff398a6a946b0be61aaa1d0eb61fd4c47708d280873b5ff29f7"} Nov 22 08:06:55 crc kubenswrapper[4735]: I1122 08:06:55.425093 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cq5w" event={"ID":"76c916a7-a1da-48e5-b7cd-1827b4d42c7a","Type":"ContainerStarted","Data":"24e4cc18ababb1f4b5b5e6721341316aae657c31e56a6fdb0c0c07b8a5fc5d7b"} Nov 22 08:06:55 crc kubenswrapper[4735]: I1122 08:06:55.441168 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h5p7s" podStartSLOduration=2.007392687 podStartE2EDuration="3.441150603s" podCreationTimestamp="2025-11-22 08:06:52 +0000 UTC" firstStartedPulling="2025-11-22 08:06:53.396068086 +0000 UTC m=+235.000406701" lastFinishedPulling="2025-11-22 08:06:54.829826012 +0000 UTC m=+236.434164617" observedRunningTime="2025-11-22 08:06:55.438762638 +0000 UTC m=+237.043101263" watchObservedRunningTime="2025-11-22 08:06:55.441150603 +0000 UTC m=+237.045489208" Nov 22 08:06:55 crc kubenswrapper[4735]: I1122 08:06:55.465047 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p8qf6" podStartSLOduration=2.396313584 podStartE2EDuration="4.465029917s" podCreationTimestamp="2025-11-22 08:06:51 +0000 UTC" firstStartedPulling="2025-11-22 08:06:52.826891923 +0000 UTC m=+234.431230528" lastFinishedPulling="2025-11-22 08:06:54.895608266 +0000 UTC m=+236.499946861" observedRunningTime="2025-11-22 08:06:55.462676914 +0000 UTC m=+237.067015539" watchObservedRunningTime="2025-11-22 08:06:55.465029917 +0000 UTC m=+237.069368532" Nov 22 08:06:55 crc kubenswrapper[4735]: I1122 08:06:55.483422 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2cq5w" podStartSLOduration=2.984810688 podStartE2EDuration="6.483400842s" podCreationTimestamp="2025-11-22 08:06:49 +0000 UTC" firstStartedPulling="2025-11-22 08:06:51.361084853 +0000 UTC m=+232.965423458" lastFinishedPulling="2025-11-22 08:06:54.859675007 +0000 UTC m=+236.464013612" observedRunningTime="2025-11-22 08:06:55.480233047 +0000 UTC m=+237.084571652" watchObservedRunningTime="2025-11-22 08:06:55.483400842 +0000 UTC m=+237.087739457" Nov 22 08:06:59 crc kubenswrapper[4735]: I1122 08:06:59.162553 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:59 crc kubenswrapper[4735]: I1122 08:06:59.162983 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:59 crc kubenswrapper[4735]: I1122 08:06:59.240003 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:06:59 crc kubenswrapper[4735]: I1122 08:06:59.493359 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l4fxx" Nov 22 08:07:00 crc kubenswrapper[4735]: I1122 08:07:00.126617 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:07:00 crc kubenswrapper[4735]: I1122 08:07:00.126685 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:07:01 crc kubenswrapper[4735]: I1122 08:07:01.179929 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2cq5w" podUID="76c916a7-a1da-48e5-b7cd-1827b4d42c7a" containerName="registry-server" probeResult="failure" output=< Nov 22 08:07:01 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:07:01 crc kubenswrapper[4735]: > Nov 22 08:07:01 crc kubenswrapper[4735]: I1122 08:07:01.501338 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:07:01 crc kubenswrapper[4735]: I1122 08:07:01.502014 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:07:01 crc kubenswrapper[4735]: I1122 08:07:01.564705 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:07:02 crc kubenswrapper[4735]: I1122 08:07:02.512961 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:07:02 crc kubenswrapper[4735]: I1122 08:07:02.513434 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:07:02 crc kubenswrapper[4735]: I1122 08:07:02.516095 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p8qf6" Nov 22 08:07:02 crc kubenswrapper[4735]: I1122 08:07:02.558866 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:07:03 crc kubenswrapper[4735]: I1122 08:07:03.512074 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h5p7s" Nov 22 08:07:08 crc kubenswrapper[4735]: I1122 08:07:08.857337 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" podUID="1f800432-2b0d-4452-902e-36d90639a132" containerName="oauth-openshift" containerID="cri-o://ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67" gracePeriod=15 Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.249342 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.284538 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6869cbc5df-mkchn"] Nov 22 08:07:09 crc kubenswrapper[4735]: E1122 08:07:09.284787 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f800432-2b0d-4452-902e-36d90639a132" containerName="oauth-openshift" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.284802 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f800432-2b0d-4452-902e-36d90639a132" containerName="oauth-openshift" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.284932 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f800432-2b0d-4452-902e-36d90639a132" containerName="oauth-openshift" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.285408 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.295380 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6869cbc5df-mkchn"] Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296744 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-error\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296780 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-session\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296799 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-provider-selection\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296844 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-audit-policies\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296880 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-serving-cert\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296907 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dfxk\" (UniqueName: \"kubernetes.io/projected/1f800432-2b0d-4452-902e-36d90639a132-kube-api-access-8dfxk\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296928 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-service-ca\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296962 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-idp-0-file-data\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.296981 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1f800432-2b0d-4452-902e-36d90639a132-audit-dir\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.297011 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-cliconfig\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.297028 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-router-certs\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.297049 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-trusted-ca-bundle\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.297065 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-login\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.297083 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-ocp-branding-template\") pod \"1f800432-2b0d-4452-902e-36d90639a132\" (UID: \"1f800432-2b0d-4452-902e-36d90639a132\") " Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.298090 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.298424 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.300254 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f800432-2b0d-4452-902e-36d90639a132-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.300904 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.301131 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.303564 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.305770 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.306954 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.307303 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.307519 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.309899 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.315420 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.315766 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.319856 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f800432-2b0d-4452-902e-36d90639a132-kube-api-access-8dfxk" (OuterVolumeSpecName: "kube-api-access-8dfxk") pod "1f800432-2b0d-4452-902e-36d90639a132" (UID: "1f800432-2b0d-4452-902e-36d90639a132"). InnerVolumeSpecName "kube-api-access-8dfxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.398877 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-service-ca\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.398927 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4nfv\" (UniqueName: \"kubernetes.io/projected/d2f2101a-d045-4eca-9851-d589a3418ebb-kube-api-access-g4nfv\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.398947 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.398966 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.398985 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-error\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399009 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-session\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399110 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399132 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399161 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399285 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399385 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d2f2101a-d045-4eca-9851-d589a3418ebb-audit-dir\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399429 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-login\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399453 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-router-certs\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399507 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-audit-policies\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399605 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399625 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399643 4735 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399656 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399669 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dfxk\" (UniqueName: \"kubernetes.io/projected/1f800432-2b0d-4452-902e-36d90639a132-kube-api-access-8dfxk\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399682 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399695 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399707 4735 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1f800432-2b0d-4452-902e-36d90639a132-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399717 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399726 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399734 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399743 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399751 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.399760 4735 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1f800432-2b0d-4452-902e-36d90639a132-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501167 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501236 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d2f2101a-d045-4eca-9851-d589a3418ebb-audit-dir\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501266 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-login\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501287 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-router-certs\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501313 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-audit-policies\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501342 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-service-ca\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501366 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4nfv\" (UniqueName: \"kubernetes.io/projected/d2f2101a-d045-4eca-9851-d589a3418ebb-kube-api-access-g4nfv\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501350 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d2f2101a-d045-4eca-9851-d589a3418ebb-audit-dir\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501390 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501418 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501441 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-error\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501485 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-session\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501532 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501559 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.501589 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.502387 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-service-ca\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.503442 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-audit-policies\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.503624 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.503873 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.504735 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-login\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.506277 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.506742 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-router-certs\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.506823 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.506932 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-user-template-error\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.507486 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" event={"ID":"1f800432-2b0d-4452-902e-36d90639a132","Type":"ContainerDied","Data":"ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67"} Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.507505 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.507539 4735 scope.go:117] "RemoveContainer" containerID="ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.507420 4735 generic.go:334] "Generic (PLEG): container finished" podID="1f800432-2b0d-4452-902e-36d90639a132" containerID="ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67" exitCode=0 Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.507743 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pc9bj" event={"ID":"1f800432-2b0d-4452-902e-36d90639a132","Type":"ContainerDied","Data":"d76588080016ba1d1ff8a49b9b7765c8c0a036161089a7b622af522a5b832fde"} Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.508036 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.508561 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.515365 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d2f2101a-d045-4eca-9851-d589a3418ebb-v4-0-config-system-session\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.524148 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4nfv\" (UniqueName: \"kubernetes.io/projected/d2f2101a-d045-4eca-9851-d589a3418ebb-kube-api-access-g4nfv\") pod \"oauth-openshift-6869cbc5df-mkchn\" (UID: \"d2f2101a-d045-4eca-9851-d589a3418ebb\") " pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.532917 4735 scope.go:117] "RemoveContainer" containerID="ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67" Nov 22 08:07:09 crc kubenswrapper[4735]: E1122 08:07:09.533477 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67\": container with ID starting with ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67 not found: ID does not exist" containerID="ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.533517 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67"} err="failed to get container status \"ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67\": rpc error: code = NotFound desc = could not find container \"ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67\": container with ID starting with ed6f394ca55e38bc9974abbb4eeec3e56367143d9f102a805df84f9086ec8f67 not found: ID does not exist" Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.544022 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pc9bj"] Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.557170 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pc9bj"] Nov 22 08:07:09 crc kubenswrapper[4735]: I1122 08:07:09.648726 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:10 crc kubenswrapper[4735]: I1122 08:07:10.060953 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6869cbc5df-mkchn"] Nov 22 08:07:10 crc kubenswrapper[4735]: W1122 08:07:10.063821 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2f2101a_d045_4eca_9851_d589a3418ebb.slice/crio-683acc21c5c8515e6c7a25520b99d3415a5613bc9ce74a2e020b40f6188577da WatchSource:0}: Error finding container 683acc21c5c8515e6c7a25520b99d3415a5613bc9ce74a2e020b40f6188577da: Status 404 returned error can't find the container with id 683acc21c5c8515e6c7a25520b99d3415a5613bc9ce74a2e020b40f6188577da Nov 22 08:07:10 crc kubenswrapper[4735]: I1122 08:07:10.180141 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:07:10 crc kubenswrapper[4735]: I1122 08:07:10.223951 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2cq5w" Nov 22 08:07:10 crc kubenswrapper[4735]: I1122 08:07:10.515049 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" event={"ID":"d2f2101a-d045-4eca-9851-d589a3418ebb","Type":"ContainerStarted","Data":"2ca74d5723a7b44d3ad3991b85565e69b4de393fbe7b902409c7b7eb2e0b01b9"} Nov 22 08:07:10 crc kubenswrapper[4735]: I1122 08:07:10.515302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" event={"ID":"d2f2101a-d045-4eca-9851-d589a3418ebb","Type":"ContainerStarted","Data":"683acc21c5c8515e6c7a25520b99d3415a5613bc9ce74a2e020b40f6188577da"} Nov 22 08:07:10 crc kubenswrapper[4735]: I1122 08:07:10.533470 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" podStartSLOduration=27.533437788 podStartE2EDuration="27.533437788s" podCreationTimestamp="2025-11-22 08:06:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:07:10.531829225 +0000 UTC m=+252.136167840" watchObservedRunningTime="2025-11-22 08:07:10.533437788 +0000 UTC m=+252.137776393" Nov 22 08:07:11 crc kubenswrapper[4735]: I1122 08:07:11.272065 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f800432-2b0d-4452-902e-36d90639a132" path="/var/lib/kubelet/pods/1f800432-2b0d-4452-902e-36d90639a132/volumes" Nov 22 08:07:11 crc kubenswrapper[4735]: I1122 08:07:11.521814 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:11 crc kubenswrapper[4735]: I1122 08:07:11.529188 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6869cbc5df-mkchn" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.171819 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq"] Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.173408 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.177813 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.178084 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.178588 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.179554 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.181096 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.188875 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq"] Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.291725 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/d031513a-6ad7-464a-8ece-fba8ff76f244-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.291966 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv2mc\" (UniqueName: \"kubernetes.io/projected/d031513a-6ad7-464a-8ece-fba8ff76f244-kube-api-access-qv2mc\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.292067 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/d031513a-6ad7-464a-8ece-fba8ff76f244-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.393739 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv2mc\" (UniqueName: \"kubernetes.io/projected/d031513a-6ad7-464a-8ece-fba8ff76f244-kube-api-access-qv2mc\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.393852 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/d031513a-6ad7-464a-8ece-fba8ff76f244-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.393964 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/d031513a-6ad7-464a-8ece-fba8ff76f244-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.395481 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/d031513a-6ad7-464a-8ece-fba8ff76f244-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.403224 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/d031513a-6ad7-464a-8ece-fba8ff76f244-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.416693 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv2mc\" (UniqueName: \"kubernetes.io/projected/d031513a-6ad7-464a-8ece-fba8ff76f244-kube-api-access-qv2mc\") pod \"cluster-monitoring-operator-6d5b84845-rw4bq\" (UID: \"d031513a-6ad7-464a-8ece-fba8ff76f244\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.494526 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" Nov 22 08:07:16 crc kubenswrapper[4735]: I1122 08:07:16.938043 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq"] Nov 22 08:07:16 crc kubenswrapper[4735]: W1122 08:07:16.946161 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd031513a_6ad7_464a_8ece_fba8ff76f244.slice/crio-020d513b6a3015309c2046f6033d1c3b060785efcd59055fd17ac0ea4c08bce6 WatchSource:0}: Error finding container 020d513b6a3015309c2046f6033d1c3b060785efcd59055fd17ac0ea4c08bce6: Status 404 returned error can't find the container with id 020d513b6a3015309c2046f6033d1c3b060785efcd59055fd17ac0ea4c08bce6 Nov 22 08:07:17 crc kubenswrapper[4735]: I1122 08:07:17.569668 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" event={"ID":"d031513a-6ad7-464a-8ece-fba8ff76f244","Type":"ContainerStarted","Data":"020d513b6a3015309c2046f6033d1c3b060785efcd59055fd17ac0ea4c08bce6"} Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.203342 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" podUID="bfd8d002-b754-4ee9-842c-48b74cb7ae8f" containerName="registry" containerID="cri-o://f00d1503aad3de00008b1477ea4e0ed604f7d39b7cc113466845254ab2306f69" gracePeriod=30 Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.579194 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" event={"ID":"d031513a-6ad7-464a-8ece-fba8ff76f244","Type":"ContainerStarted","Data":"a2f762de19d32698d73b4d94f6af5a2d9b7a42818520f6be054df127a57ac001"} Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.581732 4735 generic.go:334] "Generic (PLEG): container finished" podID="bfd8d002-b754-4ee9-842c-48b74cb7ae8f" containerID="f00d1503aad3de00008b1477ea4e0ed604f7d39b7cc113466845254ab2306f69" exitCode=0 Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.581882 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" event={"ID":"bfd8d002-b754-4ee9-842c-48b74cb7ae8f","Type":"ContainerDied","Data":"f00d1503aad3de00008b1477ea4e0ed604f7d39b7cc113466845254ab2306f69"} Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.581984 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" event={"ID":"bfd8d002-b754-4ee9-842c-48b74cb7ae8f","Type":"ContainerDied","Data":"29ad8eca9d755b4a7b7d8313f6c43f3ba729c70bc61c2dbde922ef4459dae347"} Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.582049 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29ad8eca9d755b4a7b7d8313f6c43f3ba729c70bc61c2dbde922ef4459dae347" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.584949 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.617428 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rw4bq" podStartSLOduration=1.6045740259999999 podStartE2EDuration="3.617407323s" podCreationTimestamp="2025-11-22 08:07:16 +0000 UTC" firstStartedPulling="2025-11-22 08:07:16.948620322 +0000 UTC m=+258.552958917" lastFinishedPulling="2025-11-22 08:07:18.961453569 +0000 UTC m=+260.565792214" observedRunningTime="2025-11-22 08:07:19.603164139 +0000 UTC m=+261.207502744" watchObservedRunningTime="2025-11-22 08:07:19.617407323 +0000 UTC m=+261.221745928" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.619124 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc"] Nov 22 08:07:19 crc kubenswrapper[4735]: E1122 08:07:19.619408 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfd8d002-b754-4ee9-842c-48b74cb7ae8f" containerName="registry" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.619515 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfd8d002-b754-4ee9-842c-48b74cb7ae8f" containerName="registry" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.619674 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfd8d002-b754-4ee9-842c-48b74cb7ae8f" containerName="registry" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.620113 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.623036 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-mbkqb" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.623398 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.635004 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.639887 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-certificates\") pod \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.640042 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-trusted-ca\") pod \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.640136 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-bound-sa-token\") pod \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.640241 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf2hq\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-kube-api-access-cf2hq\") pod \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.640366 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-ca-trust-extracted\") pod \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.640773 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "bfd8d002-b754-4ee9-842c-48b74cb7ae8f" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.642349 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-installation-pull-secrets\") pod \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.642454 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-tls\") pod \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\" (UID: \"bfd8d002-b754-4ee9-842c-48b74cb7ae8f\") " Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.643010 4735 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.643041 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bfd8d002-b754-4ee9-842c-48b74cb7ae8f" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.646380 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc"] Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.653378 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "bfd8d002-b754-4ee9-842c-48b74cb7ae8f" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.654124 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "bfd8d002-b754-4ee9-842c-48b74cb7ae8f" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.654227 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-kube-api-access-cf2hq" (OuterVolumeSpecName: "kube-api-access-cf2hq") pod "bfd8d002-b754-4ee9-842c-48b74cb7ae8f" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f"). InnerVolumeSpecName "kube-api-access-cf2hq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.654375 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bfd8d002-b754-4ee9-842c-48b74cb7ae8f" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.659119 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "bfd8d002-b754-4ee9-842c-48b74cb7ae8f" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.663067 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "bfd8d002-b754-4ee9-842c-48b74cb7ae8f" (UID: "bfd8d002-b754-4ee9-842c-48b74cb7ae8f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.744608 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/20b7472e-3027-4e74-a1d2-bc4e5daa3328-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-xvhdc\" (UID: \"20b7472e-3027-4e74-a1d2-bc4e5daa3328\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.745014 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.745093 4735 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.745190 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf2hq\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-kube-api-access-cf2hq\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.745423 4735 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.745517 4735 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.745643 4735 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfd8d002-b754-4ee9-842c-48b74cb7ae8f-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.847509 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/20b7472e-3027-4e74-a1d2-bc4e5daa3328-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-xvhdc\" (UID: \"20b7472e-3027-4e74-a1d2-bc4e5daa3328\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.853091 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/20b7472e-3027-4e74-a1d2-bc4e5daa3328-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-xvhdc\" (UID: \"20b7472e-3027-4e74-a1d2-bc4e5daa3328\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" Nov 22 08:07:19 crc kubenswrapper[4735]: I1122 08:07:19.939733 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" Nov 22 08:07:20 crc kubenswrapper[4735]: I1122 08:07:20.135766 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc"] Nov 22 08:07:20 crc kubenswrapper[4735]: W1122 08:07:20.154755 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20b7472e_3027_4e74_a1d2_bc4e5daa3328.slice/crio-3880dd3c201cbf7a06f8a67694d82b48cb869ef54723c4ec0cf7d65c057eb8c9 WatchSource:0}: Error finding container 3880dd3c201cbf7a06f8a67694d82b48cb869ef54723c4ec0cf7d65c057eb8c9: Status 404 returned error can't find the container with id 3880dd3c201cbf7a06f8a67694d82b48cb869ef54723c4ec0cf7d65c057eb8c9 Nov 22 08:07:20 crc kubenswrapper[4735]: I1122 08:07:20.587150 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" event={"ID":"20b7472e-3027-4e74-a1d2-bc4e5daa3328","Type":"ContainerStarted","Data":"3880dd3c201cbf7a06f8a67694d82b48cb869ef54723c4ec0cf7d65c057eb8c9"} Nov 22 08:07:20 crc kubenswrapper[4735]: I1122 08:07:20.587190 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9r74q" Nov 22 08:07:20 crc kubenswrapper[4735]: I1122 08:07:20.614822 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9r74q"] Nov 22 08:07:20 crc kubenswrapper[4735]: I1122 08:07:20.621223 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9r74q"] Nov 22 08:07:21 crc kubenswrapper[4735]: I1122 08:07:21.273300 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfd8d002-b754-4ee9-842c-48b74cb7ae8f" path="/var/lib/kubelet/pods/bfd8d002-b754-4ee9-842c-48b74cb7ae8f/volumes" Nov 22 08:07:22 crc kubenswrapper[4735]: I1122 08:07:22.600687 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" event={"ID":"20b7472e-3027-4e74-a1d2-bc4e5daa3328","Type":"ContainerStarted","Data":"4226d3de0c59bc5d146fbc29487b1d6b3accc802fb24c0239b834764f83bbcd0"} Nov 22 08:07:22 crc kubenswrapper[4735]: I1122 08:07:22.601334 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" Nov 22 08:07:22 crc kubenswrapper[4735]: I1122 08:07:22.613663 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" Nov 22 08:07:22 crc kubenswrapper[4735]: I1122 08:07:22.631018 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-xvhdc" podStartSLOduration=2.019506325 podStartE2EDuration="3.630988496s" podCreationTimestamp="2025-11-22 08:07:19 +0000 UTC" firstStartedPulling="2025-11-22 08:07:20.158116859 +0000 UTC m=+261.762455474" lastFinishedPulling="2025-11-22 08:07:21.76959903 +0000 UTC m=+263.373937645" observedRunningTime="2025-11-22 08:07:22.627326788 +0000 UTC m=+264.231665423" watchObservedRunningTime="2025-11-22 08:07:22.630988496 +0000 UTC m=+264.235327141" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.689815 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-b4l97"] Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.690940 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.694414 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.694608 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.694954 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.701985 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-68wwh" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.712275 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-b4l97"] Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.805401 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7mgp\" (UniqueName: \"kubernetes.io/projected/0aea1149-eb84-4e63-aa0d-fc367eebe68c-kube-api-access-s7mgp\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.805498 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0aea1149-eb84-4e63-aa0d-fc367eebe68c-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.805532 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0aea1149-eb84-4e63-aa0d-fc367eebe68c-metrics-client-ca\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.805586 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/0aea1149-eb84-4e63-aa0d-fc367eebe68c-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.906506 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/0aea1149-eb84-4e63-aa0d-fc367eebe68c-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.906667 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7mgp\" (UniqueName: \"kubernetes.io/projected/0aea1149-eb84-4e63-aa0d-fc367eebe68c-kube-api-access-s7mgp\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.906743 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0aea1149-eb84-4e63-aa0d-fc367eebe68c-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.906789 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0aea1149-eb84-4e63-aa0d-fc367eebe68c-metrics-client-ca\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.908495 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0aea1149-eb84-4e63-aa0d-fc367eebe68c-metrics-client-ca\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.912964 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/0aea1149-eb84-4e63-aa0d-fc367eebe68c-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.913925 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0aea1149-eb84-4e63-aa0d-fc367eebe68c-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:23 crc kubenswrapper[4735]: I1122 08:07:23.935221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7mgp\" (UniqueName: \"kubernetes.io/projected/0aea1149-eb84-4e63-aa0d-fc367eebe68c-kube-api-access-s7mgp\") pod \"prometheus-operator-db54df47d-b4l97\" (UID: \"0aea1149-eb84-4e63-aa0d-fc367eebe68c\") " pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:24 crc kubenswrapper[4735]: I1122 08:07:24.008365 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" Nov 22 08:07:24 crc kubenswrapper[4735]: I1122 08:07:24.281146 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-b4l97"] Nov 22 08:07:24 crc kubenswrapper[4735]: I1122 08:07:24.615970 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" event={"ID":"0aea1149-eb84-4e63-aa0d-fc367eebe68c","Type":"ContainerStarted","Data":"8a7ba4672b88f066a3c6082730149b0330525bb846511c36134e546b7bb02efa"} Nov 22 08:07:26 crc kubenswrapper[4735]: I1122 08:07:26.628924 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" event={"ID":"0aea1149-eb84-4e63-aa0d-fc367eebe68c","Type":"ContainerStarted","Data":"059fdb3bdc74ff0949a1dd70359e171112727db35eb1a14e24b5714a281cbad1"} Nov 22 08:07:26 crc kubenswrapper[4735]: I1122 08:07:26.629544 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" event={"ID":"0aea1149-eb84-4e63-aa0d-fc367eebe68c","Type":"ContainerStarted","Data":"6b4df12f5705f4abe9443e539ecd5b24230465810e75913e2cc8ea2564a06156"} Nov 22 08:07:26 crc kubenswrapper[4735]: I1122 08:07:26.645833 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-b4l97" podStartSLOduration=1.8887809899999999 podStartE2EDuration="3.645812917s" podCreationTimestamp="2025-11-22 08:07:23 +0000 UTC" firstStartedPulling="2025-11-22 08:07:24.290545083 +0000 UTC m=+265.894883688" lastFinishedPulling="2025-11-22 08:07:26.04757697 +0000 UTC m=+267.651915615" observedRunningTime="2025-11-22 08:07:26.642819396 +0000 UTC m=+268.247158001" watchObservedRunningTime="2025-11-22 08:07:26.645812917 +0000 UTC m=+268.250151522" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.036409 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7"] Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.037849 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.038957 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-4bz7l"] Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.039682 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.039925 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.040532 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.041246 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.041395 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.041558 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.041884 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-4fgcl" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.042987 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-fmtz8" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.065042 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-fxscm"] Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.066082 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.071781 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-6rjpt" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.071982 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.072024 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.091426 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-fxscm"] Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.126382 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7"] Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.160897 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.160953 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-sys\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.160979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c903005-b0df-467a-a3e1-3d33621ec6a1-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161032 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kprlr\" (UniqueName: \"kubernetes.io/projected/b13ad164-c900-4580-9508-e89799d9931e-kube-api-access-kprlr\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161105 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-textfile\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161163 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/1c903005-b0df-467a-a3e1-3d33621ec6a1-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161199 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161226 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161247 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-wtmp\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161278 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161304 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161338 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-root\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161353 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2hz2\" (UniqueName: \"kubernetes.io/projected/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-api-access-n2hz2\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161373 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b13ad164-c900-4580-9508-e89799d9931e-metrics-client-ca\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161539 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvbcj\" (UniqueName: \"kubernetes.io/projected/1c903005-b0df-467a-a3e1-3d33621ec6a1-kube-api-access-dvbcj\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161600 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161684 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-tls\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.161741 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/1c903005-b0df-467a-a3e1-3d33621ec6a1-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.262845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvbcj\" (UniqueName: \"kubernetes.io/projected/1c903005-b0df-467a-a3e1-3d33621ec6a1-kube-api-access-dvbcj\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.262895 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.262921 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-tls\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.262950 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/1c903005-b0df-467a-a3e1-3d33621ec6a1-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.262974 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263000 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-sys\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263018 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c903005-b0df-467a-a3e1-3d33621ec6a1-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263034 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-textfile\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263049 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kprlr\" (UniqueName: \"kubernetes.io/projected/b13ad164-c900-4580-9508-e89799d9931e-kube-api-access-kprlr\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263071 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/1c903005-b0df-467a-a3e1-3d33621ec6a1-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263087 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263104 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-wtmp\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263120 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263137 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263157 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263172 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-root\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263187 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2hz2\" (UniqueName: \"kubernetes.io/projected/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-api-access-n2hz2\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263206 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b13ad164-c900-4580-9508-e89799d9931e-metrics-client-ca\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.263898 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b13ad164-c900-4580-9508-e89799d9931e-metrics-client-ca\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.264639 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-sys\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.264753 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.265347 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-wtmp\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.265623 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/1c903005-b0df-467a-a3e1-3d33621ec6a1-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.265666 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-textfile\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.266012 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.266074 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/b13ad164-c900-4580-9508-e89799d9931e-root\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.266660 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.270972 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/1c903005-b0df-467a-a3e1-3d33621ec6a1-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.272105 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c903005-b0df-467a-a3e1-3d33621ec6a1-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.272254 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-tls\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.276113 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.286166 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b13ad164-c900-4580-9508-e89799d9931e-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.286179 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.289775 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kprlr\" (UniqueName: \"kubernetes.io/projected/b13ad164-c900-4580-9508-e89799d9931e-kube-api-access-kprlr\") pod \"node-exporter-4bz7l\" (UID: \"b13ad164-c900-4580-9508-e89799d9931e\") " pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.290270 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvbcj\" (UniqueName: \"kubernetes.io/projected/1c903005-b0df-467a-a3e1-3d33621ec6a1-kube-api-access-dvbcj\") pod \"openshift-state-metrics-566fddb674-fxscm\" (UID: \"1c903005-b0df-467a-a3e1-3d33621ec6a1\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.292846 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2hz2\" (UniqueName: \"kubernetes.io/projected/e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5-kube-api-access-n2hz2\") pod \"kube-state-metrics-777cb5bd5d-qxjr7\" (UID: \"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.352952 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.360625 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-4bz7l" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.389034 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.640419 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-4bz7l" event={"ID":"b13ad164-c900-4580-9508-e89799d9931e","Type":"ContainerStarted","Data":"d3ce5d86ce48f429fd85844dfed5ff207b830b77f4ad11a4c0fc94c7016406bf"} Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.778531 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7"] Nov 22 08:07:28 crc kubenswrapper[4735]: W1122 08:07:28.785799 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5bb74dc_6096_4c1e_b69d_9bf569b7bbe5.slice/crio-707015c5d92434b2cce81b90f6b23575e5850913862bfe9148a455d54ed2ba3a WatchSource:0}: Error finding container 707015c5d92434b2cce81b90f6b23575e5850913862bfe9148a455d54ed2ba3a: Status 404 returned error can't find the container with id 707015c5d92434b2cce81b90f6b23575e5850913862bfe9148a455d54ed2ba3a Nov 22 08:07:28 crc kubenswrapper[4735]: I1122 08:07:28.835872 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-fxscm"] Nov 22 08:07:28 crc kubenswrapper[4735]: W1122 08:07:28.841880 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c903005_b0df_467a_a3e1_3d33621ec6a1.slice/crio-25650cd1ac7b1a6ba11dc677a14628e0ae88c18d018f8fd44a1f815ac92e014e WatchSource:0}: Error finding container 25650cd1ac7b1a6ba11dc677a14628e0ae88c18d018f8fd44a1f815ac92e014e: Status 404 returned error can't find the container with id 25650cd1ac7b1a6ba11dc677a14628e0ae88c18d018f8fd44a1f815ac92e014e Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.207934 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.209793 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.211771 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.212545 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.212982 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.213209 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-5kxzb" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.213658 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.214739 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.216357 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.217513 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.227389 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.232425 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292013 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292092 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/840b334d-8b69-4a26-b454-8e2d9f51d71a-config-out\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292112 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292131 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpvp4\" (UniqueName: \"kubernetes.io/projected/840b334d-8b69-4a26-b454-8e2d9f51d71a-kube-api-access-qpvp4\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292150 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/840b334d-8b69-4a26-b454-8e2d9f51d71a-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292168 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/840b334d-8b69-4a26-b454-8e2d9f51d71a-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292185 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-web-config\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292217 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292247 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-config-volume\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292266 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/840b334d-8b69-4a26-b454-8e2d9f51d71a-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292291 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.292311 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/840b334d-8b69-4a26-b454-8e2d9f51d71a-tls-assets\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.393738 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/840b334d-8b69-4a26-b454-8e2d9f51d71a-tls-assets\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.393807 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.393885 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/840b334d-8b69-4a26-b454-8e2d9f51d71a-config-out\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.393910 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.393936 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpvp4\" (UniqueName: \"kubernetes.io/projected/840b334d-8b69-4a26-b454-8e2d9f51d71a-kube-api-access-qpvp4\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.393960 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/840b334d-8b69-4a26-b454-8e2d9f51d71a-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.393980 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/840b334d-8b69-4a26-b454-8e2d9f51d71a-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.393998 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-web-config\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.394039 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.394058 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-config-volume\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.394075 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/840b334d-8b69-4a26-b454-8e2d9f51d71a-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.394098 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.403161 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/840b334d-8b69-4a26-b454-8e2d9f51d71a-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.403402 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/840b334d-8b69-4a26-b454-8e2d9f51d71a-tls-assets\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.403528 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.403675 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/840b334d-8b69-4a26-b454-8e2d9f51d71a-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.404109 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-web-config\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.404718 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-config-volume\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.408563 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/840b334d-8b69-4a26-b454-8e2d9f51d71a-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.417751 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.418396 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.419036 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/840b334d-8b69-4a26-b454-8e2d9f51d71a-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.421264 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/840b334d-8b69-4a26-b454-8e2d9f51d71a-config-out\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.425171 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpvp4\" (UniqueName: \"kubernetes.io/projected/840b334d-8b69-4a26-b454-8e2d9f51d71a-kube-api-access-qpvp4\") pod \"alertmanager-main-0\" (UID: \"840b334d-8b69-4a26-b454-8e2d9f51d71a\") " pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.524899 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.650278 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" event={"ID":"1c903005-b0df-467a-a3e1-3d33621ec6a1","Type":"ContainerStarted","Data":"5e85b890a65b22b095df9dd8206bd1dfbe7b4bea0f8ca9ae39d7a54028560842"} Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.650770 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" event={"ID":"1c903005-b0df-467a-a3e1-3d33621ec6a1","Type":"ContainerStarted","Data":"a8a1fa7df42cd9948edb73300ad7d67111219c261b0e0ee8b1cfc1d79bd0ef63"} Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.650789 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" event={"ID":"1c903005-b0df-467a-a3e1-3d33621ec6a1","Type":"ContainerStarted","Data":"25650cd1ac7b1a6ba11dc677a14628e0ae88c18d018f8fd44a1f815ac92e014e"} Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.652635 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" event={"ID":"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5","Type":"ContainerStarted","Data":"707015c5d92434b2cce81b90f6b23575e5850913862bfe9148a455d54ed2ba3a"} Nov 22 08:07:29 crc kubenswrapper[4735]: I1122 08:07:29.952452 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Nov 22 08:07:30 crc kubenswrapper[4735]: W1122 08:07:30.009918 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod840b334d_8b69_4a26_b454_8e2d9f51d71a.slice/crio-7f2b89fbb3da663a591249d92a8bd04e6903665ea0ff8614b9e12fb35cd7090f WatchSource:0}: Error finding container 7f2b89fbb3da663a591249d92a8bd04e6903665ea0ff8614b9e12fb35cd7090f: Status 404 returned error can't find the container with id 7f2b89fbb3da663a591249d92a8bd04e6903665ea0ff8614b9e12fb35cd7090f Nov 22 08:07:30 crc kubenswrapper[4735]: I1122 08:07:30.661899 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"840b334d-8b69-4a26-b454-8e2d9f51d71a","Type":"ContainerStarted","Data":"7f2b89fbb3da663a591249d92a8bd04e6903665ea0ff8614b9e12fb35cd7090f"} Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.019693 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-8579bbc54b-z8th6"] Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.021779 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.025136 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.025400 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.025558 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.025667 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-fgp2eoal6glrn" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.025903 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.026223 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-g2qbc" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.027532 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.037422 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-8579bbc54b-z8th6"] Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.120881 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-tls\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.121108 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.121213 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-grpc-tls\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.121299 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fcf09211-515e-43bd-bc3b-f12305641d84-metrics-client-ca\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.121360 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.121386 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.121424 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.121452 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf59r\" (UniqueName: \"kubernetes.io/projected/fcf09211-515e-43bd-bc3b-f12305641d84-kube-api-access-gf59r\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.222476 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.222538 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf59r\" (UniqueName: \"kubernetes.io/projected/fcf09211-515e-43bd-bc3b-f12305641d84-kube-api-access-gf59r\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.222578 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-tls\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.222604 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.222633 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-grpc-tls\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.222685 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fcf09211-515e-43bd-bc3b-f12305641d84-metrics-client-ca\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.222731 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.222764 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.224576 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fcf09211-515e-43bd-bc3b-f12305641d84-metrics-client-ca\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.227754 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.227825 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.228742 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-tls\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.228863 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-grpc-tls\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.229879 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.235046 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/fcf09211-515e-43bd-bc3b-f12305641d84-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.248303 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf59r\" (UniqueName: \"kubernetes.io/projected/fcf09211-515e-43bd-bc3b-f12305641d84-kube-api-access-gf59r\") pod \"thanos-querier-8579bbc54b-z8th6\" (UID: \"fcf09211-515e-43bd-bc3b-f12305641d84\") " pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.339596 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.668539 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-4bz7l" event={"ID":"b13ad164-c900-4580-9508-e89799d9931e","Type":"ContainerStarted","Data":"5cab7df6a5892771a7789aa4fffef039b5e560e92b5f7ab758ec4712c95e4bf9"} Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.670205 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" event={"ID":"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5","Type":"ContainerStarted","Data":"8a507497e38130d921e1b768c9c773f7c8d07bc4c0bf1963ca843981ae7e51ae"} Nov 22 08:07:31 crc kubenswrapper[4735]: I1122 08:07:31.750371 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-8579bbc54b-z8th6"] Nov 22 08:07:31 crc kubenswrapper[4735]: W1122 08:07:31.756601 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcf09211_515e_43bd_bc3b_f12305641d84.slice/crio-c48ace2438390a4b73d35ef7fd10a33ec3b93ad93038a383aa8e1776c694bcda WatchSource:0}: Error finding container c48ace2438390a4b73d35ef7fd10a33ec3b93ad93038a383aa8e1776c694bcda: Status 404 returned error can't find the container with id c48ace2438390a4b73d35ef7fd10a33ec3b93ad93038a383aa8e1776c694bcda Nov 22 08:07:32 crc kubenswrapper[4735]: E1122 08:07:32.043068 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb13ad164_c900_4580_9508_e89799d9931e.slice/crio-conmon-5cab7df6a5892771a7789aa4fffef039b5e560e92b5f7ab758ec4712c95e4bf9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 08:07:32 crc kubenswrapper[4735]: I1122 08:07:32.680995 4735 generic.go:334] "Generic (PLEG): container finished" podID="b13ad164-c900-4580-9508-e89799d9931e" containerID="5cab7df6a5892771a7789aa4fffef039b5e560e92b5f7ab758ec4712c95e4bf9" exitCode=0 Nov 22 08:07:32 crc kubenswrapper[4735]: I1122 08:07:32.681334 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-4bz7l" event={"ID":"b13ad164-c900-4580-9508-e89799d9931e","Type":"ContainerDied","Data":"5cab7df6a5892771a7789aa4fffef039b5e560e92b5f7ab758ec4712c95e4bf9"} Nov 22 08:07:32 crc kubenswrapper[4735]: I1122 08:07:32.686062 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" event={"ID":"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5","Type":"ContainerStarted","Data":"cc934719d0de81230ac31ed9622fa4fd9c6fec0d06c85ae67569a8841add2544"} Nov 22 08:07:32 crc kubenswrapper[4735]: I1122 08:07:32.688968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" event={"ID":"fcf09211-515e-43bd-bc3b-f12305641d84","Type":"ContainerStarted","Data":"c48ace2438390a4b73d35ef7fd10a33ec3b93ad93038a383aa8e1776c694bcda"} Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:32.897764 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5ddc8ff7d8-tcm6w"] Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:32.898884 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:32.921743 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5ddc8ff7d8-tcm6w"] Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.070534 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-oauth-config\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.070682 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42qj5\" (UniqueName: \"kubernetes.io/projected/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-kube-api-access-42qj5\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.070752 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-oauth-serving-cert\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.070790 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-config\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.070807 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-serving-cert\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.070824 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-trusted-ca-bundle\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.070845 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-service-ca\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.171866 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-oauth-serving-cert\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.171943 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-serving-cert\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.171964 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-config\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.172004 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-trusted-ca-bundle\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.172033 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-service-ca\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.172061 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-oauth-config\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.172083 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42qj5\" (UniqueName: \"kubernetes.io/projected/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-kube-api-access-42qj5\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.172930 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-oauth-serving-cert\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.172930 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-config\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.173099 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-service-ca\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.173429 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-trusted-ca-bundle\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.181028 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-oauth-config\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.181251 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-serving-cert\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.188601 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42qj5\" (UniqueName: \"kubernetes.io/projected/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-kube-api-access-42qj5\") pod \"console-5ddc8ff7d8-tcm6w\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.233942 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.331744 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-857f7788fc-ltbpr"] Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.332820 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.334835 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-jjbff" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.335264 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.335804 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-9icpsgeppt5vm" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.335969 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.338599 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.340082 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.344358 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-857f7788fc-ltbpr"] Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.476586 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a846a414-147b-421d-893f-ca937978eb57-audit-log\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.476683 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx56z\" (UniqueName: \"kubernetes.io/projected/a846a414-147b-421d-893f-ca937978eb57-kube-api-access-kx56z\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.476769 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-secret-metrics-server-tls\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.476819 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a846a414-147b-421d-893f-ca937978eb57-metrics-server-audit-profiles\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.476844 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-secret-metrics-client-certs\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.476866 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a846a414-147b-421d-893f-ca937978eb57-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.476918 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-client-ca-bundle\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.578008 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a846a414-147b-421d-893f-ca937978eb57-metrics-server-audit-profiles\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.578059 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-secret-metrics-client-certs\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.578084 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a846a414-147b-421d-893f-ca937978eb57-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.578108 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-client-ca-bundle\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.578184 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a846a414-147b-421d-893f-ca937978eb57-audit-log\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.578206 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx56z\" (UniqueName: \"kubernetes.io/projected/a846a414-147b-421d-893f-ca937978eb57-kube-api-access-kx56z\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.578252 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-secret-metrics-server-tls\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.578719 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a846a414-147b-421d-893f-ca937978eb57-audit-log\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.581557 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a846a414-147b-421d-893f-ca937978eb57-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.581941 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-secret-metrics-client-certs\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.582093 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-client-ca-bundle\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.582807 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a846a414-147b-421d-893f-ca937978eb57-metrics-server-audit-profiles\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.590422 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a846a414-147b-421d-893f-ca937978eb57-secret-metrics-server-tls\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.597104 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx56z\" (UniqueName: \"kubernetes.io/projected/a846a414-147b-421d-893f-ca937978eb57-kube-api-access-kx56z\") pod \"metrics-server-857f7788fc-ltbpr\" (UID: \"a846a414-147b-421d-893f-ca937978eb57\") " pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.661271 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.699105 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" event={"ID":"1c903005-b0df-467a-a3e1-3d33621ec6a1","Type":"ContainerStarted","Data":"01788d050500e86b78f42fcecc4731c71aeb912fb5f90ee07dd6abd53294cfa1"} Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.716834 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-fxscm" podStartSLOduration=2.236531788 podStartE2EDuration="5.716815849s" podCreationTimestamp="2025-11-22 08:07:28 +0000 UTC" firstStartedPulling="2025-11-22 08:07:29.195586528 +0000 UTC m=+270.799925133" lastFinishedPulling="2025-11-22 08:07:32.675870579 +0000 UTC m=+274.280209194" observedRunningTime="2025-11-22 08:07:33.714586128 +0000 UTC m=+275.318924753" watchObservedRunningTime="2025-11-22 08:07:33.716815849 +0000 UTC m=+275.321154454" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.730030 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" event={"ID":"e5bb74dc-6096-4c1e-b69d-9bf569b7bbe5","Type":"ContainerStarted","Data":"76f06fc59f10338644bfa1b83952d8bcae59be6c6adba176d99ca676b1058a9e"} Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.735936 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-4bz7l" event={"ID":"b13ad164-c900-4580-9508-e89799d9931e","Type":"ContainerStarted","Data":"e152d05a08fa772367338ae1b8b56099f303703e59a0b13bb62c90d6ae687948"} Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.845332 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-qxjr7" podStartSLOduration=3.386874848 podStartE2EDuration="5.845317195s" podCreationTimestamp="2025-11-22 08:07:28 +0000 UTC" firstStartedPulling="2025-11-22 08:07:28.789232846 +0000 UTC m=+270.393571461" lastFinishedPulling="2025-11-22 08:07:31.247675203 +0000 UTC m=+272.852013808" observedRunningTime="2025-11-22 08:07:33.752516712 +0000 UTC m=+275.356855317" watchObservedRunningTime="2025-11-22 08:07:33.845317195 +0000 UTC m=+275.449655800" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.847336 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4"] Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.848014 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.852702 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.853024 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.876923 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4"] Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.881500 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5ddc8ff7d8-tcm6w"] Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.882228 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/cd22293a-9bc6-49cf-b2f6-ccecb7f52a1a-monitoring-plugin-cert\") pod \"monitoring-plugin-78b44f74f7-kxkr4\" (UID: \"cd22293a-9bc6-49cf-b2f6-ccecb7f52a1a\") " pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.984699 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/cd22293a-9bc6-49cf-b2f6-ccecb7f52a1a-monitoring-plugin-cert\") pod \"monitoring-plugin-78b44f74f7-kxkr4\" (UID: \"cd22293a-9bc6-49cf-b2f6-ccecb7f52a1a\") " pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" Nov 22 08:07:33 crc kubenswrapper[4735]: I1122 08:07:33.989526 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/cd22293a-9bc6-49cf-b2f6-ccecb7f52a1a-monitoring-plugin-cert\") pod \"monitoring-plugin-78b44f74f7-kxkr4\" (UID: \"cd22293a-9bc6-49cf-b2f6-ccecb7f52a1a\") " pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.142441 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-857f7788fc-ltbpr"] Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.211643 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.296498 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.309925 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.314620 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.315076 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-29fup38hc3nhg" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.315205 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.316684 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.317238 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-67ncw" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.317507 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.317560 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.317620 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.317683 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.318140 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.318316 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.328986 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.335402 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.336063 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490550 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-web-config\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490599 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490627 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76hc5\" (UniqueName: \"kubernetes.io/projected/a6c75463-22e2-416d-a878-d2998e83c460-kube-api-access-76hc5\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490659 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490689 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-config\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490706 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490745 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490767 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a6c75463-22e2-416d-a878-d2998e83c460-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490788 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490812 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490833 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490851 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490877 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490897 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490933 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a6c75463-22e2-416d-a878-d2998e83c460-config-out\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490955 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490973 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.490992 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592192 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592257 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592299 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a6c75463-22e2-416d-a878-d2998e83c460-config-out\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592324 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592345 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592378 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592428 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-web-config\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592450 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592489 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76hc5\" (UniqueName: \"kubernetes.io/projected/a6c75463-22e2-416d-a878-d2998e83c460-kube-api-access-76hc5\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592516 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592547 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-config\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592571 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592613 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592637 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a6c75463-22e2-416d-a878-d2998e83c460-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592662 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592686 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592712 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.592735 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.597016 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.597576 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.601732 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.601939 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.602775 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.604401 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.606000 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a6c75463-22e2-416d-a878-d2998e83c460-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.610939 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-config\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.615017 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.615941 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.616060 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.618336 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a6c75463-22e2-416d-a878-d2998e83c460-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.622892 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a6c75463-22e2-416d-a878-d2998e83c460-config-out\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.623510 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.623904 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.634021 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.634858 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a6c75463-22e2-416d-a878-d2998e83c460-web-config\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.638341 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76hc5\" (UniqueName: \"kubernetes.io/projected/a6c75463-22e2-416d-a878-d2998e83c460-kube-api-access-76hc5\") pod \"prometheus-k8s-0\" (UID: \"a6c75463-22e2-416d-a878-d2998e83c460\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.745857 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5ddc8ff7d8-tcm6w" event={"ID":"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1","Type":"ContainerStarted","Data":"fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad"} Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.745909 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5ddc8ff7d8-tcm6w" event={"ID":"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1","Type":"ContainerStarted","Data":"c2817aaf684c69bd89f9127dd37727653d70c2f80c5c6cc72fbc0a7dc06744c6"} Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.750171 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-4bz7l" event={"ID":"b13ad164-c900-4580-9508-e89799d9931e","Type":"ContainerStarted","Data":"81c3c0fab41796e12b72f4d114ed32a37abedc5ddee4c008a2ba264fb9e1e1ed"} Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.755162 4735 generic.go:334] "Generic (PLEG): container finished" podID="840b334d-8b69-4a26-b454-8e2d9f51d71a" containerID="376d180839411c267fee8b8e5936b0e280c1169d7fec5e955954ab34cf885fc8" exitCode=0 Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.756157 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"840b334d-8b69-4a26-b454-8e2d9f51d71a","Type":"ContainerDied","Data":"376d180839411c267fee8b8e5936b0e280c1169d7fec5e955954ab34cf885fc8"} Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.770523 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5ddc8ff7d8-tcm6w" podStartSLOduration=2.770500672 podStartE2EDuration="2.770500672s" podCreationTimestamp="2025-11-22 08:07:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:07:34.761654623 +0000 UTC m=+276.365993228" watchObservedRunningTime="2025-11-22 08:07:34.770500672 +0000 UTC m=+276.374839267" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.781701 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-4bz7l" podStartSLOduration=4.528812992 podStartE2EDuration="6.781684813s" podCreationTimestamp="2025-11-22 08:07:28 +0000 UTC" firstStartedPulling="2025-11-22 08:07:28.379875354 +0000 UTC m=+269.984213959" lastFinishedPulling="2025-11-22 08:07:30.632747175 +0000 UTC m=+272.237085780" observedRunningTime="2025-11-22 08:07:34.780540963 +0000 UTC m=+276.384879578" watchObservedRunningTime="2025-11-22 08:07:34.781684813 +0000 UTC m=+276.386023418" Nov 22 08:07:34 crc kubenswrapper[4735]: I1122 08:07:34.935983 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.196578 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4"] Nov 22 08:07:35 crc kubenswrapper[4735]: W1122 08:07:35.409542 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6c75463_22e2_416d_a878_d2998e83c460.slice/crio-f2c1e47bf20a9bdd2e065dc874395bd75b7b0d6b37a876c408910e3b0a08dfec WatchSource:0}: Error finding container f2c1e47bf20a9bdd2e065dc874395bd75b7b0d6b37a876c408910e3b0a08dfec: Status 404 returned error can't find the container with id f2c1e47bf20a9bdd2e065dc874395bd75b7b0d6b37a876c408910e3b0a08dfec Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.410503 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.762352 4735 generic.go:334] "Generic (PLEG): container finished" podID="a6c75463-22e2-416d-a878-d2998e83c460" containerID="00666cd60575a8eb8ba4ac6d188a9ecbbbda1fd2394f8e8bccb812a2c6fbba96" exitCode=0 Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.762396 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"a6c75463-22e2-416d-a878-d2998e83c460","Type":"ContainerDied","Data":"00666cd60575a8eb8ba4ac6d188a9ecbbbda1fd2394f8e8bccb812a2c6fbba96"} Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.762436 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"a6c75463-22e2-416d-a878-d2998e83c460","Type":"ContainerStarted","Data":"f2c1e47bf20a9bdd2e065dc874395bd75b7b0d6b37a876c408910e3b0a08dfec"} Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.764239 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" event={"ID":"fcf09211-515e-43bd-bc3b-f12305641d84","Type":"ContainerStarted","Data":"cab022d88b21a7f8bb24d95063113c609732a4d83470a96d1cc922b1c8ba1fdf"} Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.764276 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" event={"ID":"fcf09211-515e-43bd-bc3b-f12305641d84","Type":"ContainerStarted","Data":"6e7431874e85cb4a3b7eb0964bc88b23c8b4e2de5ed18a1e03663186f9b75616"} Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.764288 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" event={"ID":"fcf09211-515e-43bd-bc3b-f12305641d84","Type":"ContainerStarted","Data":"3fb07c6d8952c1386feda4fb2f2a2bf26797b31d27d1027e052d11d3288cdb14"} Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.765943 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" event={"ID":"a846a414-147b-421d-893f-ca937978eb57","Type":"ContainerStarted","Data":"7a1ae6c88b4b0dfcd161ec22a2507ff08a8c0c3b06c62217b2bc090938326d97"} Nov 22 08:07:35 crc kubenswrapper[4735]: I1122 08:07:35.767061 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" event={"ID":"cd22293a-9bc6-49cf-b2f6-ccecb7f52a1a","Type":"ContainerStarted","Data":"99239288e2b9f546de29ade0963379cec431f3e5c011e028874577d24e05f600"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.780760 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" event={"ID":"a846a414-147b-421d-893f-ca937978eb57","Type":"ContainerStarted","Data":"0507379cf24fc4a15aad61c2f8660ffb2c7b781f55b6360ca0e6cbee4d4765dd"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.784933 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"840b334d-8b69-4a26-b454-8e2d9f51d71a","Type":"ContainerStarted","Data":"3c1b96b7a01db0c01b5a7e5fb3fa85351c7038516bedc9f99e481ecae3d53a74"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.784974 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"840b334d-8b69-4a26-b454-8e2d9f51d71a","Type":"ContainerStarted","Data":"3db4e925b4b2de96ab2780c399adbcc17ef6ae32b4c39062f965d20d2f4b5099"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.784989 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"840b334d-8b69-4a26-b454-8e2d9f51d71a","Type":"ContainerStarted","Data":"258e2336bee2241687e9bdcf417830520568035ce77cb4aeb4a6f0d9ee7779ca"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.786330 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" event={"ID":"cd22293a-9bc6-49cf-b2f6-ccecb7f52a1a","Type":"ContainerStarted","Data":"f2bf856bf93558e74537fe4de4694c3cf80c977b345c363ad7b032e633d07697"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.786494 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.792893 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" event={"ID":"fcf09211-515e-43bd-bc3b-f12305641d84","Type":"ContainerStarted","Data":"3e8beb310ecb4c80d0dbb27b08dec6e9337a70e9bc67f37aa7a2cb26369e896a"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.792966 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" event={"ID":"fcf09211-515e-43bd-bc3b-f12305641d84","Type":"ContainerStarted","Data":"90cd8980102729488ec512b937643f40e63b91997219b1cbf1866acc3c5ed6e9"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.792983 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" event={"ID":"fcf09211-515e-43bd-bc3b-f12305641d84","Type":"ContainerStarted","Data":"75ac7c3675d16d6ccada37f8a16734e16179e9af4f143137b99f889edfbaabeb"} Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.793094 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.802025 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.803172 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" podStartSLOduration=2.494229265 podStartE2EDuration="4.803150048s" podCreationTimestamp="2025-11-22 08:07:33 +0000 UTC" firstStartedPulling="2025-11-22 08:07:34.979673504 +0000 UTC m=+276.584012109" lastFinishedPulling="2025-11-22 08:07:37.288594287 +0000 UTC m=+278.892932892" observedRunningTime="2025-11-22 08:07:37.797236559 +0000 UTC m=+279.401575174" watchObservedRunningTime="2025-11-22 08:07:37.803150048 +0000 UTC m=+279.407488653" Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.829976 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" podStartSLOduration=2.226598689 podStartE2EDuration="7.829955261s" podCreationTimestamp="2025-11-22 08:07:30 +0000 UTC" firstStartedPulling="2025-11-22 08:07:31.759175021 +0000 UTC m=+273.363513626" lastFinishedPulling="2025-11-22 08:07:37.362531593 +0000 UTC m=+278.966870198" observedRunningTime="2025-11-22 08:07:37.82582467 +0000 UTC m=+279.430163295" watchObservedRunningTime="2025-11-22 08:07:37.829955261 +0000 UTC m=+279.434293866" Nov 22 08:07:37 crc kubenswrapper[4735]: I1122 08:07:37.865922 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-78b44f74f7-kxkr4" podStartSLOduration=2.790194628 podStartE2EDuration="4.86589757s" podCreationTimestamp="2025-11-22 08:07:33 +0000 UTC" firstStartedPulling="2025-11-22 08:07:35.204738075 +0000 UTC m=+276.809076680" lastFinishedPulling="2025-11-22 08:07:37.280441017 +0000 UTC m=+278.884779622" observedRunningTime="2025-11-22 08:07:37.844669028 +0000 UTC m=+279.449007643" watchObservedRunningTime="2025-11-22 08:07:37.86589757 +0000 UTC m=+279.470236175" Nov 22 08:07:38 crc kubenswrapper[4735]: I1122 08:07:38.803291 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"840b334d-8b69-4a26-b454-8e2d9f51d71a","Type":"ContainerStarted","Data":"cf16b182ef2ad2bf95bf37ab2131cdfbe3e70c8bf952418f15a6defb8264d94f"} Nov 22 08:07:38 crc kubenswrapper[4735]: I1122 08:07:38.803638 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"840b334d-8b69-4a26-b454-8e2d9f51d71a","Type":"ContainerStarted","Data":"876c1a3485ef311ef2c553b67206059b46f5847d9f9142736e13e8f906d36c0b"} Nov 22 08:07:38 crc kubenswrapper[4735]: I1122 08:07:38.803654 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"840b334d-8b69-4a26-b454-8e2d9f51d71a","Type":"ContainerStarted","Data":"2197e60d1710a0cd18c2e8866f80a86d76b6f715b1a331a070d38cdd2c228b73"} Nov 22 08:07:38 crc kubenswrapper[4735]: I1122 08:07:38.846663 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.578358203 podStartE2EDuration="9.846613855s" podCreationTimestamp="2025-11-22 08:07:29 +0000 UTC" firstStartedPulling="2025-11-22 08:07:30.012186966 +0000 UTC m=+271.616525571" lastFinishedPulling="2025-11-22 08:07:37.280442618 +0000 UTC m=+278.884781223" observedRunningTime="2025-11-22 08:07:38.843432169 +0000 UTC m=+280.447770814" watchObservedRunningTime="2025-11-22 08:07:38.846613855 +0000 UTC m=+280.450952470" Nov 22 08:07:39 crc kubenswrapper[4735]: I1122 08:07:39.813125 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"a6c75463-22e2-416d-a878-d2998e83c460","Type":"ContainerStarted","Data":"45e87d7a257c90e13ae9290996f00a9b1bd75f39fb047ae9ff1e313936a4c830"} Nov 22 08:07:39 crc kubenswrapper[4735]: I1122 08:07:39.813426 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"a6c75463-22e2-416d-a878-d2998e83c460","Type":"ContainerStarted","Data":"38ae6fa9112062d0af7e9d28a248d5fa1c692567ea13b3d1e3f05471d052ef7b"} Nov 22 08:07:40 crc kubenswrapper[4735]: I1122 08:07:40.823141 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"a6c75463-22e2-416d-a878-d2998e83c460","Type":"ContainerStarted","Data":"a665cca6932d9d0a94ce3e96795ffc44fff518b75e08107b770c8e057f6b8f2d"} Nov 22 08:07:40 crc kubenswrapper[4735]: I1122 08:07:40.823198 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"a6c75463-22e2-416d-a878-d2998e83c460","Type":"ContainerStarted","Data":"d9490f8f8ae4fd0fe1b12a679b04fea2f240792c6f3dbf35d45bf67a23d5c4df"} Nov 22 08:07:40 crc kubenswrapper[4735]: I1122 08:07:40.823213 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"a6c75463-22e2-416d-a878-d2998e83c460","Type":"ContainerStarted","Data":"a4d9a224f1a15684fe1687bfe2e9e5238360ca2c466ff198a9f63da6b20b94f8"} Nov 22 08:07:40 crc kubenswrapper[4735]: I1122 08:07:40.823225 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"a6c75463-22e2-416d-a878-d2998e83c460","Type":"ContainerStarted","Data":"d3bc74f7034c9c63e4262934db3b2017d19ef4735935284f92586dc389fdc520"} Nov 22 08:07:40 crc kubenswrapper[4735]: I1122 08:07:40.880043 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=3.134967083 podStartE2EDuration="6.880017586s" podCreationTimestamp="2025-11-22 08:07:34 +0000 UTC" firstStartedPulling="2025-11-22 08:07:35.764347111 +0000 UTC m=+277.368685716" lastFinishedPulling="2025-11-22 08:07:39.509397604 +0000 UTC m=+281.113736219" observedRunningTime="2025-11-22 08:07:40.873901252 +0000 UTC m=+282.478239937" watchObservedRunningTime="2025-11-22 08:07:40.880017586 +0000 UTC m=+282.484356191" Nov 22 08:07:41 crc kubenswrapper[4735]: I1122 08:07:41.361761 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" Nov 22 08:07:43 crc kubenswrapper[4735]: I1122 08:07:43.235145 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:43 crc kubenswrapper[4735]: I1122 08:07:43.235244 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:43 crc kubenswrapper[4735]: I1122 08:07:43.242929 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:43 crc kubenswrapper[4735]: I1122 08:07:43.860425 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:07:43 crc kubenswrapper[4735]: I1122 08:07:43.936716 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sjvm9"] Nov 22 08:07:44 crc kubenswrapper[4735]: I1122 08:07:44.936623 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:07:53 crc kubenswrapper[4735]: I1122 08:07:53.661851 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:07:53 crc kubenswrapper[4735]: I1122 08:07:53.662395 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:08:08 crc kubenswrapper[4735]: I1122 08:08:08.993591 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-sjvm9" podUID="d417a7ee-7687-4086-8047-39e259a9bfbe" containerName="console" containerID="cri-o://fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b" gracePeriod=15 Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.368316 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sjvm9_d417a7ee-7687-4086-8047-39e259a9bfbe/console/0.log" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.368913 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.528611 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-service-ca\") pod \"d417a7ee-7687-4086-8047-39e259a9bfbe\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.529042 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-oauth-config\") pod \"d417a7ee-7687-4086-8047-39e259a9bfbe\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.529075 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-oauth-serving-cert\") pod \"d417a7ee-7687-4086-8047-39e259a9bfbe\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.529092 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h58m\" (UniqueName: \"kubernetes.io/projected/d417a7ee-7687-4086-8047-39e259a9bfbe-kube-api-access-6h58m\") pod \"d417a7ee-7687-4086-8047-39e259a9bfbe\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.529114 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-trusted-ca-bundle\") pod \"d417a7ee-7687-4086-8047-39e259a9bfbe\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.529162 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-serving-cert\") pod \"d417a7ee-7687-4086-8047-39e259a9bfbe\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.529180 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-console-config\") pod \"d417a7ee-7687-4086-8047-39e259a9bfbe\" (UID: \"d417a7ee-7687-4086-8047-39e259a9bfbe\") " Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.529388 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-service-ca" (OuterVolumeSpecName: "service-ca") pod "d417a7ee-7687-4086-8047-39e259a9bfbe" (UID: "d417a7ee-7687-4086-8047-39e259a9bfbe"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.529523 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.530113 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d417a7ee-7687-4086-8047-39e259a9bfbe" (UID: "d417a7ee-7687-4086-8047-39e259a9bfbe"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.530366 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d417a7ee-7687-4086-8047-39e259a9bfbe" (UID: "d417a7ee-7687-4086-8047-39e259a9bfbe"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.530531 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-console-config" (OuterVolumeSpecName: "console-config") pod "d417a7ee-7687-4086-8047-39e259a9bfbe" (UID: "d417a7ee-7687-4086-8047-39e259a9bfbe"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.538518 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d417a7ee-7687-4086-8047-39e259a9bfbe-kube-api-access-6h58m" (OuterVolumeSpecName: "kube-api-access-6h58m") pod "d417a7ee-7687-4086-8047-39e259a9bfbe" (UID: "d417a7ee-7687-4086-8047-39e259a9bfbe"). InnerVolumeSpecName "kube-api-access-6h58m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.539151 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d417a7ee-7687-4086-8047-39e259a9bfbe" (UID: "d417a7ee-7687-4086-8047-39e259a9bfbe"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.539638 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d417a7ee-7687-4086-8047-39e259a9bfbe" (UID: "d417a7ee-7687-4086-8047-39e259a9bfbe"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.630399 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.630454 4735 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.630515 4735 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.630524 4735 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d417a7ee-7687-4086-8047-39e259a9bfbe-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.630534 4735 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d417a7ee-7687-4086-8047-39e259a9bfbe-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:09 crc kubenswrapper[4735]: I1122 08:08:09.630544 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h58m\" (UniqueName: \"kubernetes.io/projected/d417a7ee-7687-4086-8047-39e259a9bfbe-kube-api-access-6h58m\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.040558 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sjvm9_d417a7ee-7687-4086-8047-39e259a9bfbe/console/0.log" Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.040633 4735 generic.go:334] "Generic (PLEG): container finished" podID="d417a7ee-7687-4086-8047-39e259a9bfbe" containerID="fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b" exitCode=2 Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.040683 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sjvm9" event={"ID":"d417a7ee-7687-4086-8047-39e259a9bfbe","Type":"ContainerDied","Data":"fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b"} Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.040727 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sjvm9" Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.040764 4735 scope.go:117] "RemoveContainer" containerID="fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b" Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.040744 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sjvm9" event={"ID":"d417a7ee-7687-4086-8047-39e259a9bfbe","Type":"ContainerDied","Data":"ef3b3445da7cdc478758902e9bb13787348f4acb5cf03b8afcdc356c391b8b8e"} Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.069182 4735 scope.go:117] "RemoveContainer" containerID="fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b" Nov 22 08:08:10 crc kubenswrapper[4735]: E1122 08:08:10.069875 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b\": container with ID starting with fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b not found: ID does not exist" containerID="fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b" Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.069936 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b"} err="failed to get container status \"fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b\": rpc error: code = NotFound desc = could not find container \"fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b\": container with ID starting with fe4896385afa7bca86c49ab928364662fd59166643b46e19a1729d788f7b597b not found: ID does not exist" Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.078719 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sjvm9"] Nov 22 08:08:10 crc kubenswrapper[4735]: I1122 08:08:10.081800 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-sjvm9"] Nov 22 08:08:11 crc kubenswrapper[4735]: I1122 08:08:11.274345 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d417a7ee-7687-4086-8047-39e259a9bfbe" path="/var/lib/kubelet/pods/d417a7ee-7687-4086-8047-39e259a9bfbe/volumes" Nov 22 08:08:13 crc kubenswrapper[4735]: I1122 08:08:13.671150 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:08:13 crc kubenswrapper[4735]: I1122 08:08:13.677090 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-857f7788fc-ltbpr" Nov 22 08:08:34 crc kubenswrapper[4735]: I1122 08:08:34.937001 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:08:34 crc kubenswrapper[4735]: I1122 08:08:34.981332 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:08:35 crc kubenswrapper[4735]: I1122 08:08:35.243637 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Nov 22 08:08:46 crc kubenswrapper[4735]: I1122 08:08:46.131287 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:08:46 crc kubenswrapper[4735]: I1122 08:08:46.131905 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.497615 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-56b67fb7dd-w5qgb"] Nov 22 08:08:54 crc kubenswrapper[4735]: E1122 08:08:54.498174 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d417a7ee-7687-4086-8047-39e259a9bfbe" containerName="console" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.498186 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d417a7ee-7687-4086-8047-39e259a9bfbe" containerName="console" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.498300 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d417a7ee-7687-4086-8047-39e259a9bfbe" containerName="console" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.498718 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.521997 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-56b67fb7dd-w5qgb"] Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.536605 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-oauth-serving-cert\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.536667 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-oauth-config\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.536696 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz9sq\" (UniqueName: \"kubernetes.io/projected/4c2b6c8e-3c56-4e20-a28b-182129e41afd-kube-api-access-kz9sq\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.536718 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-service-ca\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.536885 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-config\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.536989 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-serving-cert\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.537042 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-trusted-ca-bundle\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.639143 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-oauth-config\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.639199 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-service-ca\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.639216 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz9sq\" (UniqueName: \"kubernetes.io/projected/4c2b6c8e-3c56-4e20-a28b-182129e41afd-kube-api-access-kz9sq\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.639253 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-config\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.639277 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-serving-cert\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.639295 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-trusted-ca-bundle\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.639362 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-oauth-serving-cert\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.640561 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-service-ca\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.640593 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-config\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.640635 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-oauth-serving-cert\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.640726 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-trusted-ca-bundle\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.645914 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-serving-cert\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.647631 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-oauth-config\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.655400 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz9sq\" (UniqueName: \"kubernetes.io/projected/4c2b6c8e-3c56-4e20-a28b-182129e41afd-kube-api-access-kz9sq\") pod \"console-56b67fb7dd-w5qgb\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:54 crc kubenswrapper[4735]: I1122 08:08:54.817295 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:08:55 crc kubenswrapper[4735]: I1122 08:08:55.001846 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-56b67fb7dd-w5qgb"] Nov 22 08:08:55 crc kubenswrapper[4735]: I1122 08:08:55.339284 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-56b67fb7dd-w5qgb" event={"ID":"4c2b6c8e-3c56-4e20-a28b-182129e41afd","Type":"ContainerStarted","Data":"c86af4eaaadec71e246e5cda1aa5bca586583afdd74b0fee27518b1aa9384c33"} Nov 22 08:08:55 crc kubenswrapper[4735]: I1122 08:08:55.339950 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-56b67fb7dd-w5qgb" event={"ID":"4c2b6c8e-3c56-4e20-a28b-182129e41afd","Type":"ContainerStarted","Data":"cedcbd5dd91e4f2a0afa57e85f953ee407428f4cce3f5536ca5fcd594ca69b3f"} Nov 22 08:08:55 crc kubenswrapper[4735]: I1122 08:08:55.383236 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-56b67fb7dd-w5qgb" podStartSLOduration=1.383207583 podStartE2EDuration="1.383207583s" podCreationTimestamp="2025-11-22 08:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:08:55.377898749 +0000 UTC m=+356.982237354" watchObservedRunningTime="2025-11-22 08:08:55.383207583 +0000 UTC m=+356.987546208" Nov 22 08:09:04 crc kubenswrapper[4735]: I1122 08:09:04.818396 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:09:04 crc kubenswrapper[4735]: I1122 08:09:04.819023 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:09:04 crc kubenswrapper[4735]: I1122 08:09:04.823683 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:09:05 crc kubenswrapper[4735]: I1122 08:09:05.408879 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:09:05 crc kubenswrapper[4735]: I1122 08:09:05.468096 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5ddc8ff7d8-tcm6w"] Nov 22 08:09:16 crc kubenswrapper[4735]: I1122 08:09:16.131321 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:09:16 crc kubenswrapper[4735]: I1122 08:09:16.131871 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.519941 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-5ddc8ff7d8-tcm6w" podUID="41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" containerName="console" containerID="cri-o://fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad" gracePeriod=15 Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.909805 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5ddc8ff7d8-tcm6w_41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1/console/0.log" Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.910175 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.996541 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-serving-cert\") pod \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.996675 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-config\") pod \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.996743 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-oauth-serving-cert\") pod \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.996865 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42qj5\" (UniqueName: \"kubernetes.io/projected/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-kube-api-access-42qj5\") pod \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.996911 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-trusted-ca-bundle\") pod \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.996949 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-oauth-config\") pod \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.997007 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-service-ca\") pod \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\" (UID: \"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1\") " Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.997395 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-config" (OuterVolumeSpecName: "console-config") pod "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" (UID: "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.997660 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" (UID: "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.997893 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" (UID: "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:09:30 crc kubenswrapper[4735]: I1122 08:09:30.997976 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-service-ca" (OuterVolumeSpecName: "service-ca") pod "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" (UID: "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.002631 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-kube-api-access-42qj5" (OuterVolumeSpecName: "kube-api-access-42qj5") pod "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" (UID: "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1"). InnerVolumeSpecName "kube-api-access-42qj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.002911 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" (UID: "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.004112 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" (UID: "41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.099117 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.099155 4735 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.099169 4735 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.099222 4735 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.099235 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42qj5\" (UniqueName: \"kubernetes.io/projected/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-kube-api-access-42qj5\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.099243 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.099271 4735 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.604351 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5ddc8ff7d8-tcm6w_41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1/console/0.log" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.604425 4735 generic.go:334] "Generic (PLEG): container finished" podID="41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" containerID="fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad" exitCode=2 Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.604506 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5ddc8ff7d8-tcm6w" event={"ID":"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1","Type":"ContainerDied","Data":"fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad"} Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.604512 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5ddc8ff7d8-tcm6w" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.604548 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5ddc8ff7d8-tcm6w" event={"ID":"41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1","Type":"ContainerDied","Data":"c2817aaf684c69bd89f9127dd37727653d70c2f80c5c6cc72fbc0a7dc06744c6"} Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.604574 4735 scope.go:117] "RemoveContainer" containerID="fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.624729 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5ddc8ff7d8-tcm6w"] Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.630743 4735 scope.go:117] "RemoveContainer" containerID="fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad" Nov 22 08:09:31 crc kubenswrapper[4735]: E1122 08:09:31.631533 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad\": container with ID starting with fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad not found: ID does not exist" containerID="fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.631591 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad"} err="failed to get container status \"fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad\": rpc error: code = NotFound desc = could not find container \"fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad\": container with ID starting with fc6da9e75d8d0ad74e75bc8e90f70f3fbb49fd24c09e4c4a4b69a40c4dcee2ad not found: ID does not exist" Nov 22 08:09:31 crc kubenswrapper[4735]: I1122 08:09:31.637435 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5ddc8ff7d8-tcm6w"] Nov 22 08:09:33 crc kubenswrapper[4735]: I1122 08:09:33.273737 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" path="/var/lib/kubelet/pods/41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1/volumes" Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.131757 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.132692 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.132756 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.134016 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c483bef07756ce17b03463d41a2cb38d389b0b284cac0a615aac582d5c89aec3"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.134186 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://c483bef07756ce17b03463d41a2cb38d389b0b284cac0a615aac582d5c89aec3" gracePeriod=600 Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.721415 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="c483bef07756ce17b03463d41a2cb38d389b0b284cac0a615aac582d5c89aec3" exitCode=0 Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.721452 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"c483bef07756ce17b03463d41a2cb38d389b0b284cac0a615aac582d5c89aec3"} Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.721782 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"f7aa357304ffacda92a92578afeb23951ae04837d67816d3ca4f766ea770f256"} Nov 22 08:09:46 crc kubenswrapper[4735]: I1122 08:09:46.721804 4735 scope.go:117] "RemoveContainer" containerID="305ca36d357afb098f26e45ee80517694bfc637364b72bc427885f3584aab213" Nov 22 08:11:46 crc kubenswrapper[4735]: I1122 08:11:46.131937 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:11:46 crc kubenswrapper[4735]: I1122 08:11:46.132657 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:11:59 crc kubenswrapper[4735]: I1122 08:11:59.451267 4735 scope.go:117] "RemoveContainer" containerID="f00d1503aad3de00008b1477ea4e0ed604f7d39b7cc113466845254ab2306f69" Nov 22 08:12:16 crc kubenswrapper[4735]: I1122 08:12:16.131811 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:12:16 crc kubenswrapper[4735]: I1122 08:12:16.132671 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:12:46 crc kubenswrapper[4735]: I1122 08:12:46.131885 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:12:46 crc kubenswrapper[4735]: I1122 08:12:46.132627 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:12:46 crc kubenswrapper[4735]: I1122 08:12:46.132716 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:12:46 crc kubenswrapper[4735]: I1122 08:12:46.133653 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f7aa357304ffacda92a92578afeb23951ae04837d67816d3ca4f766ea770f256"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:12:46 crc kubenswrapper[4735]: I1122 08:12:46.133785 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://f7aa357304ffacda92a92578afeb23951ae04837d67816d3ca4f766ea770f256" gracePeriod=600 Nov 22 08:12:46 crc kubenswrapper[4735]: I1122 08:12:46.421138 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="f7aa357304ffacda92a92578afeb23951ae04837d67816d3ca4f766ea770f256" exitCode=0 Nov 22 08:12:46 crc kubenswrapper[4735]: I1122 08:12:46.421284 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"f7aa357304ffacda92a92578afeb23951ae04837d67816d3ca4f766ea770f256"} Nov 22 08:12:46 crc kubenswrapper[4735]: I1122 08:12:46.421451 4735 scope.go:117] "RemoveContainer" containerID="c483bef07756ce17b03463d41a2cb38d389b0b284cac0a615aac582d5c89aec3" Nov 22 08:12:47 crc kubenswrapper[4735]: I1122 08:12:47.432356 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"a3aab057b98bc5c57ec42d5f2111ef56b0a63d358e465d259bfc8b9b235ad36f"} Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.028727 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8"] Nov 22 08:12:51 crc kubenswrapper[4735]: E1122 08:12:51.029537 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" containerName="console" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.029576 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" containerName="console" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.029694 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ac43ac-7b33-4b3a-abaf-016bc2cb0ee1" containerName="console" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.030624 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.032643 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.037986 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8"] Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.213387 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.213442 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.213694 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zbnr\" (UniqueName: \"kubernetes.io/projected/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-kube-api-access-4zbnr\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.314612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.314681 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.314752 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zbnr\" (UniqueName: \"kubernetes.io/projected/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-kube-api-access-4zbnr\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.315496 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.315528 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.336201 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zbnr\" (UniqueName: \"kubernetes.io/projected/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-kube-api-access-4zbnr\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.353195 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:51 crc kubenswrapper[4735]: I1122 08:12:51.581714 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8"] Nov 22 08:12:52 crc kubenswrapper[4735]: I1122 08:12:52.469637 4735 generic.go:334] "Generic (PLEG): container finished" podID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerID="705a1db41f9da69a797f9a48024889d3313d5f8d7a54de505e7c01e2169d4fa1" exitCode=0 Nov 22 08:12:52 crc kubenswrapper[4735]: I1122 08:12:52.469720 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" event={"ID":"b6c8e68c-78c8-4d48-be93-a5b5d23204e0","Type":"ContainerDied","Data":"705a1db41f9da69a797f9a48024889d3313d5f8d7a54de505e7c01e2169d4fa1"} Nov 22 08:12:52 crc kubenswrapper[4735]: I1122 08:12:52.469772 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" event={"ID":"b6c8e68c-78c8-4d48-be93-a5b5d23204e0","Type":"ContainerStarted","Data":"b937a7cb8e13307a07702cc2308640c8403187e78015aa6ec3f9fe5fc287397d"} Nov 22 08:12:52 crc kubenswrapper[4735]: I1122 08:12:52.473360 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:12:54 crc kubenswrapper[4735]: I1122 08:12:54.487434 4735 generic.go:334] "Generic (PLEG): container finished" podID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerID="02730d133b5cacbbcc5e64f4de3c9ad71dee6e2958aa5e12f524ba33fa25c8c7" exitCode=0 Nov 22 08:12:54 crc kubenswrapper[4735]: I1122 08:12:54.487566 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" event={"ID":"b6c8e68c-78c8-4d48-be93-a5b5d23204e0","Type":"ContainerDied","Data":"02730d133b5cacbbcc5e64f4de3c9ad71dee6e2958aa5e12f524ba33fa25c8c7"} Nov 22 08:12:55 crc kubenswrapper[4735]: I1122 08:12:55.500060 4735 generic.go:334] "Generic (PLEG): container finished" podID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerID="c0ade630c664c418e9fc929302218606235363de967c10fef2059f61df94ae30" exitCode=0 Nov 22 08:12:55 crc kubenswrapper[4735]: I1122 08:12:55.500165 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" event={"ID":"b6c8e68c-78c8-4d48-be93-a5b5d23204e0","Type":"ContainerDied","Data":"c0ade630c664c418e9fc929302218606235363de967c10fef2059f61df94ae30"} Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.764094 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.894544 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-util\") pod \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.894603 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-bundle\") pod \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.894639 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zbnr\" (UniqueName: \"kubernetes.io/projected/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-kube-api-access-4zbnr\") pod \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\" (UID: \"b6c8e68c-78c8-4d48-be93-a5b5d23204e0\") " Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.896973 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-bundle" (OuterVolumeSpecName: "bundle") pod "b6c8e68c-78c8-4d48-be93-a5b5d23204e0" (UID: "b6c8e68c-78c8-4d48-be93-a5b5d23204e0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.909698 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-kube-api-access-4zbnr" (OuterVolumeSpecName: "kube-api-access-4zbnr") pod "b6c8e68c-78c8-4d48-be93-a5b5d23204e0" (UID: "b6c8e68c-78c8-4d48-be93-a5b5d23204e0"). InnerVolumeSpecName "kube-api-access-4zbnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.915517 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-util" (OuterVolumeSpecName: "util") pod "b6c8e68c-78c8-4d48-be93-a5b5d23204e0" (UID: "b6c8e68c-78c8-4d48-be93-a5b5d23204e0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.995841 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zbnr\" (UniqueName: \"kubernetes.io/projected/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-kube-api-access-4zbnr\") on node \"crc\" DevicePath \"\"" Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.995890 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:12:56 crc kubenswrapper[4735]: I1122 08:12:56.995904 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b6c8e68c-78c8-4d48-be93-a5b5d23204e0-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:12:57 crc kubenswrapper[4735]: I1122 08:12:57.514909 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" event={"ID":"b6c8e68c-78c8-4d48-be93-a5b5d23204e0","Type":"ContainerDied","Data":"b937a7cb8e13307a07702cc2308640c8403187e78015aa6ec3f9fe5fc287397d"} Nov 22 08:12:57 crc kubenswrapper[4735]: I1122 08:12:57.514976 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b937a7cb8e13307a07702cc2308640c8403187e78015aa6ec3f9fe5fc287397d" Nov 22 08:12:57 crc kubenswrapper[4735]: I1122 08:12:57.515004 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8" Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.234049 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pl875"] Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.235102 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovn-controller" containerID="cri-o://1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b" gracePeriod=30 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.235176 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="northd" containerID="cri-o://332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f" gracePeriod=30 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.235305 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovn-acl-logging" containerID="cri-o://344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307" gracePeriod=30 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.235335 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kube-rbac-proxy-node" containerID="cri-o://ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6" gracePeriod=30 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.235199 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="sbdb" containerID="cri-o://b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6" gracePeriod=30 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.235409 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770" gracePeriod=30 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.235446 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="nbdb" containerID="cri-o://b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d" gracePeriod=30 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.277841 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" containerID="cri-o://f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848" gracePeriod=30 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.558718 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovnkube-controller/3.log" Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.561226 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovn-acl-logging/0.log" Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.561879 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovn-controller/0.log" Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562851 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848" exitCode=0 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562875 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d" exitCode=0 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562884 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f" exitCode=0 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562892 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307" exitCode=143 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562900 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b" exitCode=143 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562936 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848"} Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562961 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d"} Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562972 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f"} Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562981 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307"} Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.562991 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b"} Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.563007 4735 scope.go:117] "RemoveContainer" containerID="c8414c7748da0836b179423b00151d34b83dd7ef2be64b038b4665b8057009d9" Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.565101 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/2.log" Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.566407 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/1.log" Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.566438 4735 generic.go:334] "Generic (PLEG): container finished" podID="61dbab7f-c903-4b55-9d90-aacfd12f7f0f" containerID="8f71d0c8bea2f1c5a42c8365198e3fe40631e28bfc7c5f174a032e124a354f1a" exitCode=2 Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.566460 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-85z2h" event={"ID":"61dbab7f-c903-4b55-9d90-aacfd12f7f0f","Type":"ContainerDied","Data":"8f71d0c8bea2f1c5a42c8365198e3fe40631e28bfc7c5f174a032e124a354f1a"} Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.566886 4735 scope.go:117] "RemoveContainer" containerID="8f71d0c8bea2f1c5a42c8365198e3fe40631e28bfc7c5f174a032e124a354f1a" Nov 22 08:13:02 crc kubenswrapper[4735]: E1122 08:13:02.567042 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-85z2h_openshift-multus(61dbab7f-c903-4b55-9d90-aacfd12f7f0f)\"" pod="openshift-multus/multus-85z2h" podUID="61dbab7f-c903-4b55-9d90-aacfd12f7f0f" Nov 22 08:13:02 crc kubenswrapper[4735]: I1122 08:13:02.594045 4735 scope.go:117] "RemoveContainer" containerID="a946f35649eccb94951ebbf6a56a37079c530cbad1cf644ad76cec1fd5450d73" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.447212 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovn-acl-logging/0.log" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.448623 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovn-controller/0.log" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.449031 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523161 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c4ppq"] Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523412 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovn-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523433 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovn-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523445 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523457 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523487 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="sbdb" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523494 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="sbdb" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523503 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523510 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523522 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerName="pull" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523529 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerName="pull" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523537 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kube-rbac-proxy-node" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523544 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kube-rbac-proxy-node" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523552 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="nbdb" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523561 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="nbdb" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523571 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523577 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523583 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerName="util" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523590 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerName="util" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523599 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovn-acl-logging" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523605 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovn-acl-logging" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523611 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523616 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523626 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="northd" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523631 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="northd" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523637 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerName="extract" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523642 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerName="extract" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523650 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kubecfg-setup" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523655 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kubecfg-setup" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523769 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523778 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523785 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523792 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="nbdb" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523800 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523806 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="kube-rbac-proxy-node" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523816 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="sbdb" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523823 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c8e68c-78c8-4d48-be93-a5b5d23204e0" containerName="extract" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523832 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="northd" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523839 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523847 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovn-acl-logging" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523854 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovn-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.523949 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.523956 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.524050 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.524145 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.524152 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2719040d-3088-41c7-8f16-5508d78669e2" containerName="ovnkube-controller" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.525798 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.575582 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovn-acl-logging/0.log" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576114 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pl875_2719040d-3088-41c7-8f16-5508d78669e2/ovn-controller/0.log" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576639 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6" exitCode=0 Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576682 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770" exitCode=0 Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576692 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719040d-3088-41c7-8f16-5508d78669e2" containerID="ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6" exitCode=0 Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576732 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576800 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6"} Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576853 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770"} Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576865 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6"} Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576874 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pl875" event={"ID":"2719040d-3088-41c7-8f16-5508d78669e2","Type":"ContainerDied","Data":"1f0b9a794a58e1b354d17e739c2c69508e496ff0bab41382b9a2c8a4ac474046"} Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.576889 4735 scope.go:117] "RemoveContainer" containerID="f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.578519 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/2.log" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.594872 4735 scope.go:117] "RemoveContainer" containerID="b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601111 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2719040d-3088-41c7-8f16-5508d78669e2-ovn-node-metrics-cert\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601354 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvx5h\" (UniqueName: \"kubernetes.io/projected/2719040d-3088-41c7-8f16-5508d78669e2-kube-api-access-dvx5h\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601449 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-var-lib-openvswitch\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601597 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-systemd-units\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601691 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-script-lib\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601785 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-systemd\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601884 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-kubelet\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601980 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-node-log\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602082 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-openvswitch\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602183 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-ovn-kubernetes\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602269 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-log-socket\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602368 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-env-overrides\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602457 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-etc-openvswitch\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602577 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-netns\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601526 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601684 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.601912 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602021 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-node-log" (OuterVolumeSpecName: "node-log") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602085 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602693 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602118 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602264 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602300 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-log-socket" (OuterVolumeSpecName: "log-socket") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602730 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602749 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.602858 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603099 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-netd\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603244 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-bin\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603325 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603416 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-config\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603264 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603560 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603638 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-slash\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603727 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-ovn\") pod \"2719040d-3088-41c7-8f16-5508d78669e2\" (UID: \"2719040d-3088-41c7-8f16-5508d78669e2\") " Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603853 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-slash" (OuterVolumeSpecName: "host-slash") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603907 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.603962 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.604138 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-var-lib-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.604260 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-cni-bin\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.604352 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.604446 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovnkube-config\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.608764 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-node-log\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.608890 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovnkube-script-lib\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.608985 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-log-socket\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609114 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-cni-netd\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609256 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-systemd\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609369 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-systemd-units\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609447 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-env-overrides\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609559 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-etc-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609650 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovn-node-metrics-cert\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609735 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-kubelet\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609829 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-slash\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.609955 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g56pp\" (UniqueName: \"kubernetes.io/projected/1aa1f126-12a1-47a0-909b-e2dea899eedd-kube-api-access-g56pp\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.617829 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.617934 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618045 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-run-netns\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618128 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-ovn\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618294 4735 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618361 4735 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618422 4735 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618532 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618608 4735 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618675 4735 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-node-log\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618742 4735 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618807 4735 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618881 4735 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-log-socket\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.618946 4735 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.619011 4735 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.619077 4735 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.619139 4735 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.619206 4735 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.619279 4735 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.619347 4735 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2719040d-3088-41c7-8f16-5508d78669e2-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.619417 4735 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-host-slash\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.621803 4735 scope.go:117] "RemoveContainer" containerID="b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.622337 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2719040d-3088-41c7-8f16-5508d78669e2-kube-api-access-dvx5h" (OuterVolumeSpecName: "kube-api-access-dvx5h") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "kube-api-access-dvx5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.630055 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2719040d-3088-41c7-8f16-5508d78669e2-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.669549 4735 scope.go:117] "RemoveContainer" containerID="332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.674321 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "2719040d-3088-41c7-8f16-5508d78669e2" (UID: "2719040d-3088-41c7-8f16-5508d78669e2"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.694983 4735 scope.go:117] "RemoveContainer" containerID="11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.717663 4735 scope.go:117] "RemoveContainer" containerID="ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.722334 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.722602 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.722693 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-run-netns\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.722848 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-ovn\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.722948 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-var-lib-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723053 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-cni-bin\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723139 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723231 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovnkube-config\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723324 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-node-log\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723404 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovnkube-script-lib\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723497 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-log-socket\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723586 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-cni-netd\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723678 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-systemd\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723760 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-systemd-units\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723840 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-etc-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723909 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-env-overrides\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.723986 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovn-node-metrics-cert\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724060 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-kubelet\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724131 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-slash\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724219 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g56pp\" (UniqueName: \"kubernetes.io/projected/1aa1f126-12a1-47a0-909b-e2dea899eedd-kube-api-access-g56pp\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724326 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2719040d-3088-41c7-8f16-5508d78669e2-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724395 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvx5h\" (UniqueName: \"kubernetes.io/projected/2719040d-3088-41c7-8f16-5508d78669e2-kube-api-access-dvx5h\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724460 4735 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2719040d-3088-41c7-8f16-5508d78669e2-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724827 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-node-log\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724897 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-var-lib-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.722707 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724870 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-ovn\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.722440 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.724830 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovnkube-config\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.722748 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-run-netns\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725027 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-cni-bin\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725058 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725092 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-etc-openvswitch\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725123 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-log-socket\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725155 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-cni-netd\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725186 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-run-systemd\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725218 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-systemd-units\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725321 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-kubelet\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725365 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1aa1f126-12a1-47a0-909b-e2dea899eedd-host-slash\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.725769 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-env-overrides\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.726257 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovnkube-script-lib\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.732101 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1aa1f126-12a1-47a0-909b-e2dea899eedd-ovn-node-metrics-cert\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.740072 4735 scope.go:117] "RemoveContainer" containerID="344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.752360 4735 scope.go:117] "RemoveContainer" containerID="1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.771361 4735 scope.go:117] "RemoveContainer" containerID="c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.776801 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g56pp\" (UniqueName: \"kubernetes.io/projected/1aa1f126-12a1-47a0-909b-e2dea899eedd-kube-api-access-g56pp\") pod \"ovnkube-node-c4ppq\" (UID: \"1aa1f126-12a1-47a0-909b-e2dea899eedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.794414 4735 scope.go:117] "RemoveContainer" containerID="f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.798595 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848\": container with ID starting with f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848 not found: ID does not exist" containerID="f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.798629 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848"} err="failed to get container status \"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848\": rpc error: code = NotFound desc = could not find container \"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848\": container with ID starting with f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.798653 4735 scope.go:117] "RemoveContainer" containerID="b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.804600 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\": container with ID starting with b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6 not found: ID does not exist" containerID="b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.804648 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6"} err="failed to get container status \"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\": rpc error: code = NotFound desc = could not find container \"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\": container with ID starting with b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.804673 4735 scope.go:117] "RemoveContainer" containerID="b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.808546 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\": container with ID starting with b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d not found: ID does not exist" containerID="b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.808575 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d"} err="failed to get container status \"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\": rpc error: code = NotFound desc = could not find container \"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\": container with ID starting with b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.808593 4735 scope.go:117] "RemoveContainer" containerID="332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.809459 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\": container with ID starting with 332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f not found: ID does not exist" containerID="332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.809497 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f"} err="failed to get container status \"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\": rpc error: code = NotFound desc = could not find container \"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\": container with ID starting with 332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.809515 4735 scope.go:117] "RemoveContainer" containerID="11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.813621 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\": container with ID starting with 11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770 not found: ID does not exist" containerID="11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.813651 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770"} err="failed to get container status \"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\": rpc error: code = NotFound desc = could not find container \"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\": container with ID starting with 11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.813671 4735 scope.go:117] "RemoveContainer" containerID="ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.817091 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\": container with ID starting with ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6 not found: ID does not exist" containerID="ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.817125 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6"} err="failed to get container status \"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\": rpc error: code = NotFound desc = could not find container \"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\": container with ID starting with ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.817147 4735 scope.go:117] "RemoveContainer" containerID="344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.820699 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\": container with ID starting with 344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307 not found: ID does not exist" containerID="344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.820723 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307"} err="failed to get container status \"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\": rpc error: code = NotFound desc = could not find container \"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\": container with ID starting with 344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.820744 4735 scope.go:117] "RemoveContainer" containerID="1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.822378 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\": container with ID starting with 1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b not found: ID does not exist" containerID="1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.822408 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b"} err="failed to get container status \"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\": rpc error: code = NotFound desc = could not find container \"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\": container with ID starting with 1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.822449 4735 scope.go:117] "RemoveContainer" containerID="c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156" Nov 22 08:13:03 crc kubenswrapper[4735]: E1122 08:13:03.825754 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\": container with ID starting with c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156 not found: ID does not exist" containerID="c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.825780 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156"} err="failed to get container status \"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\": rpc error: code = NotFound desc = could not find container \"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\": container with ID starting with c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.825797 4735 scope.go:117] "RemoveContainer" containerID="f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.826003 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848"} err="failed to get container status \"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848\": rpc error: code = NotFound desc = could not find container \"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848\": container with ID starting with f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.826018 4735 scope.go:117] "RemoveContainer" containerID="b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.826206 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6"} err="failed to get container status \"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\": rpc error: code = NotFound desc = could not find container \"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\": container with ID starting with b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.826224 4735 scope.go:117] "RemoveContainer" containerID="b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.826405 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d"} err="failed to get container status \"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\": rpc error: code = NotFound desc = could not find container \"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\": container with ID starting with b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.826420 4735 scope.go:117] "RemoveContainer" containerID="332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.828894 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f"} err="failed to get container status \"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\": rpc error: code = NotFound desc = could not find container \"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\": container with ID starting with 332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.828915 4735 scope.go:117] "RemoveContainer" containerID="11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.830755 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770"} err="failed to get container status \"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\": rpc error: code = NotFound desc = could not find container \"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\": container with ID starting with 11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.830775 4735 scope.go:117] "RemoveContainer" containerID="ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.834542 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6"} err="failed to get container status \"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\": rpc error: code = NotFound desc = could not find container \"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\": container with ID starting with ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.834563 4735 scope.go:117] "RemoveContainer" containerID="344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.838919 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307"} err="failed to get container status \"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\": rpc error: code = NotFound desc = could not find container \"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\": container with ID starting with 344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.838946 4735 scope.go:117] "RemoveContainer" containerID="1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.839172 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b"} err="failed to get container status \"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\": rpc error: code = NotFound desc = could not find container \"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\": container with ID starting with 1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.839196 4735 scope.go:117] "RemoveContainer" containerID="c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.839369 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156"} err="failed to get container status \"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\": rpc error: code = NotFound desc = could not find container \"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\": container with ID starting with c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.839389 4735 scope.go:117] "RemoveContainer" containerID="f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.839642 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.848469 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848"} err="failed to get container status \"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848\": rpc error: code = NotFound desc = could not find container \"f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848\": container with ID starting with f9046f205a32e85fe96c3b7543ba8b431e2074bf18bdbecce06b3c2f8030c848 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.848508 4735 scope.go:117] "RemoveContainer" containerID="b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.851715 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6"} err="failed to get container status \"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\": rpc error: code = NotFound desc = could not find container \"b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6\": container with ID starting with b67266ccbeb761b4c8c805f931afe7dfd0f81d534c68e8a03b8ebc55bca9bcf6 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.851757 4735 scope.go:117] "RemoveContainer" containerID="b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.855058 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d"} err="failed to get container status \"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\": rpc error: code = NotFound desc = could not find container \"b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d\": container with ID starting with b081569acaf2697ea1b4d0576f7fda9b5dca4779616271d81387931d5c211c8d not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.855089 4735 scope.go:117] "RemoveContainer" containerID="332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.864070 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f"} err="failed to get container status \"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\": rpc error: code = NotFound desc = could not find container \"332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f\": container with ID starting with 332fb83ccc0991361b9baa7903fc21633ffca3db0f99c210a610bf451695b31f not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.864109 4735 scope.go:117] "RemoveContainer" containerID="11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.867393 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770"} err="failed to get container status \"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\": rpc error: code = NotFound desc = could not find container \"11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770\": container with ID starting with 11d011b8c41dc43ec7a82e22694c3bc70539df5c937ab4ed76d1eb0eae1a7770 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.867416 4735 scope.go:117] "RemoveContainer" containerID="ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.873244 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6"} err="failed to get container status \"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\": rpc error: code = NotFound desc = could not find container \"ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6\": container with ID starting with ed6ed77d4b5100aa80f400d510d992fda0b60f737e174ca15b3d80552d2b6db6 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.873290 4735 scope.go:117] "RemoveContainer" containerID="344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.874050 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307"} err="failed to get container status \"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\": rpc error: code = NotFound desc = could not find container \"344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307\": container with ID starting with 344f40116713511e6659e8c7220cd44e4d0d125a0add60d889c54b270bb77307 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.874071 4735 scope.go:117] "RemoveContainer" containerID="1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.874863 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b"} err="failed to get container status \"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\": rpc error: code = NotFound desc = could not find container \"1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b\": container with ID starting with 1a8f4d5f37f7ef83407b8393d7268efd53567cda718f9a8adca5532e39e8ff9b not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.874892 4735 scope.go:117] "RemoveContainer" containerID="c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.875336 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156"} err="failed to get container status \"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\": rpc error: code = NotFound desc = could not find container \"c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156\": container with ID starting with c35d5f39ff0fc4dd484134c95d9f9c644e2f2c6a8d63c087e3a52faeb038d156 not found: ID does not exist" Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.952037 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pl875"] Nov 22 08:13:03 crc kubenswrapper[4735]: I1122 08:13:03.965423 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pl875"] Nov 22 08:13:04 crc kubenswrapper[4735]: I1122 08:13:04.586763 4735 generic.go:334] "Generic (PLEG): container finished" podID="1aa1f126-12a1-47a0-909b-e2dea899eedd" containerID="4e49ecf9d7d2dd46622f7dfd0c1dc85a5cd8cb505f00bac565115064e3275417" exitCode=0 Nov 22 08:13:04 crc kubenswrapper[4735]: I1122 08:13:04.586815 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerDied","Data":"4e49ecf9d7d2dd46622f7dfd0c1dc85a5cd8cb505f00bac565115064e3275417"} Nov 22 08:13:04 crc kubenswrapper[4735]: I1122 08:13:04.586853 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"5b0d208cd3c274711870e129fcfdd33fbc551cb4d1451ab22bfb9dca906abc47"} Nov 22 08:13:05 crc kubenswrapper[4735]: I1122 08:13:05.277343 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2719040d-3088-41c7-8f16-5508d78669e2" path="/var/lib/kubelet/pods/2719040d-3088-41c7-8f16-5508d78669e2/volumes" Nov 22 08:13:05 crc kubenswrapper[4735]: I1122 08:13:05.595765 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"745eb1110885a30777883612a6e7e3d9dbdb3418cf1f09f21e7b796539f67ef2"} Nov 22 08:13:05 crc kubenswrapper[4735]: I1122 08:13:05.595806 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"67ae3f8bf78d2dba565aadb3c2fd38275b11cd4dcea60eb9e18c36a26e6a6fa8"} Nov 22 08:13:05 crc kubenswrapper[4735]: I1122 08:13:05.595816 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"a9f7992865845fe250cb88d12f3f490c066efe9dc30703535243edc08dcb37c2"} Nov 22 08:13:05 crc kubenswrapper[4735]: I1122 08:13:05.595825 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"27f9a2b830f7dffd5fa2ff8ff46b02371758b06c9cdf94128077e902862f8ce6"} Nov 22 08:13:05 crc kubenswrapper[4735]: I1122 08:13:05.595833 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"83305d463744b7a5997781ad5813e2577afb2d9f15e40e863e8cdd1798c13fcd"} Nov 22 08:13:05 crc kubenswrapper[4735]: I1122 08:13:05.595840 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"fa733a36a825d2d6a0e90b86874743c9b613c5214b7699b48132352b6ca74894"} Nov 22 08:13:08 crc kubenswrapper[4735]: I1122 08:13:08.614034 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"1d3c7827c23aed9d02f8f4f0c3f1700ee66da2ae41fe12974fa10b2d454e10d9"} Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.419528 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm"] Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.420402 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.422096 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.422599 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.423041 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-8c5n8" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.469217 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579"] Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.469930 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.471473 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.476814 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-95kmm" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.481233 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz"] Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.482098 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.534160 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50c1fd51-0497-44e6-b9d6-52b48e5a4133-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579\" (UID: \"50c1fd51-0497-44e6-b9d6-52b48e5a4133\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.534244 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57655\" (UniqueName: \"kubernetes.io/projected/eab1ae65-df97-4ef5-9876-dccbac6d5304-kube-api-access-57655\") pod \"obo-prometheus-operator-668cf9dfbb-sxnpm\" (UID: \"eab1ae65-df97-4ef5-9876-dccbac6d5304\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.534294 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50c1fd51-0497-44e6-b9d6-52b48e5a4133-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579\" (UID: \"50c1fd51-0497-44e6-b9d6-52b48e5a4133\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.534349 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73f91b2d-3de6-40da-ac7b-e97c7b63a8ad-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz\" (UID: \"73f91b2d-3de6-40da-ac7b-e97c7b63a8ad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.534371 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73f91b2d-3de6-40da-ac7b-e97c7b63a8ad-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz\" (UID: \"73f91b2d-3de6-40da-ac7b-e97c7b63a8ad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.635190 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73f91b2d-3de6-40da-ac7b-e97c7b63a8ad-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz\" (UID: \"73f91b2d-3de6-40da-ac7b-e97c7b63a8ad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.635235 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73f91b2d-3de6-40da-ac7b-e97c7b63a8ad-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz\" (UID: \"73f91b2d-3de6-40da-ac7b-e97c7b63a8ad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.635272 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50c1fd51-0497-44e6-b9d6-52b48e5a4133-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579\" (UID: \"50c1fd51-0497-44e6-b9d6-52b48e5a4133\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.635317 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57655\" (UniqueName: \"kubernetes.io/projected/eab1ae65-df97-4ef5-9876-dccbac6d5304-kube-api-access-57655\") pod \"obo-prometheus-operator-668cf9dfbb-sxnpm\" (UID: \"eab1ae65-df97-4ef5-9876-dccbac6d5304\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.635344 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50c1fd51-0497-44e6-b9d6-52b48e5a4133-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579\" (UID: \"50c1fd51-0497-44e6-b9d6-52b48e5a4133\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.644122 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73f91b2d-3de6-40da-ac7b-e97c7b63a8ad-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz\" (UID: \"73f91b2d-3de6-40da-ac7b-e97c7b63a8ad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.644207 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50c1fd51-0497-44e6-b9d6-52b48e5a4133-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579\" (UID: \"50c1fd51-0497-44e6-b9d6-52b48e5a4133\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.646377 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73f91b2d-3de6-40da-ac7b-e97c7b63a8ad-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz\" (UID: \"73f91b2d-3de6-40da-ac7b-e97c7b63a8ad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.652004 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50c1fd51-0497-44e6-b9d6-52b48e5a4133-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579\" (UID: \"50c1fd51-0497-44e6-b9d6-52b48e5a4133\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.652407 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5wrtk"] Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.654573 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57655\" (UniqueName: \"kubernetes.io/projected/eab1ae65-df97-4ef5-9876-dccbac6d5304-kube-api-access-57655\") pod \"obo-prometheus-operator-668cf9dfbb-sxnpm\" (UID: \"eab1ae65-df97-4ef5-9876-dccbac6d5304\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.656581 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.668942 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-rwb78" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.669151 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.736342 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8mjp\" (UniqueName: \"kubernetes.io/projected/93273ee2-9156-4b65-abe9-1e020aa4ea55-kube-api-access-x8mjp\") pod \"observability-operator-d8bb48f5d-5wrtk\" (UID: \"93273ee2-9156-4b65-abe9-1e020aa4ea55\") " pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.736446 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/93273ee2-9156-4b65-abe9-1e020aa4ea55-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5wrtk\" (UID: \"93273ee2-9156-4b65-abe9-1e020aa4ea55\") " pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.739106 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.770605 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(4043d1d173d08f14a4d0152e8ad430750bbd60ff28c27133081e53df6be3ead7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.770727 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(4043d1d173d08f14a4d0152e8ad430750bbd60ff28c27133081e53df6be3ead7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.770752 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(4043d1d173d08f14a4d0152e8ad430750bbd60ff28c27133081e53df6be3ead7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.770810 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators(eab1ae65-df97-4ef5-9876-dccbac6d5304)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators(eab1ae65-df97-4ef5-9876-dccbac6d5304)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(4043d1d173d08f14a4d0152e8ad430750bbd60ff28c27133081e53df6be3ead7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" podUID="eab1ae65-df97-4ef5-9876-dccbac6d5304" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.790468 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.799071 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.831230 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(3a3932162844610f83fec0d4bbfd5205f2ba6e979fd2c0aa5e5e492719a37afd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.831271 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(3a3932162844610f83fec0d4bbfd5205f2ba6e979fd2c0aa5e5e492719a37afd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.831293 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(3a3932162844610f83fec0d4bbfd5205f2ba6e979fd2c0aa5e5e492719a37afd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.831329 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators(50c1fd51-0497-44e6-b9d6-52b48e5a4133)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators(50c1fd51-0497-44e6-b9d6-52b48e5a4133)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(3a3932162844610f83fec0d4bbfd5205f2ba6e979fd2c0aa5e5e492719a37afd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" podUID="50c1fd51-0497-44e6-b9d6-52b48e5a4133" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.837431 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/93273ee2-9156-4b65-abe9-1e020aa4ea55-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5wrtk\" (UID: \"93273ee2-9156-4b65-abe9-1e020aa4ea55\") " pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.837576 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8mjp\" (UniqueName: \"kubernetes.io/projected/93273ee2-9156-4b65-abe9-1e020aa4ea55-kube-api-access-x8mjp\") pod \"observability-operator-d8bb48f5d-5wrtk\" (UID: \"93273ee2-9156-4b65-abe9-1e020aa4ea55\") " pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.842390 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/93273ee2-9156-4b65-abe9-1e020aa4ea55-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5wrtk\" (UID: \"93273ee2-9156-4b65-abe9-1e020aa4ea55\") " pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.848945 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(63079140fefa27ea0bc6c9f689f6639faf1cf1273c2627b06a533a63411a77b6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.849015 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(63079140fefa27ea0bc6c9f689f6639faf1cf1273c2627b06a533a63411a77b6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.849038 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(63079140fefa27ea0bc6c9f689f6639faf1cf1273c2627b06a533a63411a77b6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:09 crc kubenswrapper[4735]: E1122 08:13:09.849102 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators(73f91b2d-3de6-40da-ac7b-e97c7b63a8ad)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators(73f91b2d-3de6-40da-ac7b-e97c7b63a8ad)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(63079140fefa27ea0bc6c9f689f6639faf1cf1273c2627b06a533a63411a77b6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" podUID="73f91b2d-3de6-40da-ac7b-e97c7b63a8ad" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.862154 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-jq2p9"] Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.862954 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.865167 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8mjp\" (UniqueName: \"kubernetes.io/projected/93273ee2-9156-4b65-abe9-1e020aa4ea55-kube-api-access-x8mjp\") pod \"observability-operator-d8bb48f5d-5wrtk\" (UID: \"93273ee2-9156-4b65-abe9-1e020aa4ea55\") " pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.865987 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-m5nvg" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.939129 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc5129c5-6a4c-4d09-807f-1811604f3193-openshift-service-ca\") pod \"perses-operator-5446b9c989-jq2p9\" (UID: \"fc5129c5-6a4c-4d09-807f-1811604f3193\") " pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:09 crc kubenswrapper[4735]: I1122 08:13:09.939509 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wswd\" (UniqueName: \"kubernetes.io/projected/fc5129c5-6a4c-4d09-807f-1811604f3193-kube-api-access-2wswd\") pod \"perses-operator-5446b9c989-jq2p9\" (UID: \"fc5129c5-6a4c-4d09-807f-1811604f3193\") " pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.008151 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.031018 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(04eff4dde254200fc0989e478378c75604f67d62fe0f7bbdf97908b8e185eddd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.031084 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(04eff4dde254200fc0989e478378c75604f67d62fe0f7bbdf97908b8e185eddd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.031106 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(04eff4dde254200fc0989e478378c75604f67d62fe0f7bbdf97908b8e185eddd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.031154 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-5wrtk_openshift-operators(93273ee2-9156-4b65-abe9-1e020aa4ea55)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-5wrtk_openshift-operators(93273ee2-9156-4b65-abe9-1e020aa4ea55)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(04eff4dde254200fc0989e478378c75604f67d62fe0f7bbdf97908b8e185eddd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" podUID="93273ee2-9156-4b65-abe9-1e020aa4ea55" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.040267 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wswd\" (UniqueName: \"kubernetes.io/projected/fc5129c5-6a4c-4d09-807f-1811604f3193-kube-api-access-2wswd\") pod \"perses-operator-5446b9c989-jq2p9\" (UID: \"fc5129c5-6a4c-4d09-807f-1811604f3193\") " pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.040359 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc5129c5-6a4c-4d09-807f-1811604f3193-openshift-service-ca\") pod \"perses-operator-5446b9c989-jq2p9\" (UID: \"fc5129c5-6a4c-4d09-807f-1811604f3193\") " pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.041209 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc5129c5-6a4c-4d09-807f-1811604f3193-openshift-service-ca\") pod \"perses-operator-5446b9c989-jq2p9\" (UID: \"fc5129c5-6a4c-4d09-807f-1811604f3193\") " pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.065017 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wswd\" (UniqueName: \"kubernetes.io/projected/fc5129c5-6a4c-4d09-807f-1811604f3193-kube-api-access-2wswd\") pod \"perses-operator-5446b9c989-jq2p9\" (UID: \"fc5129c5-6a4c-4d09-807f-1811604f3193\") " pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.177615 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.205638 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(7807384f4402a511d28829969e54fa91f8df902d1affc369202fa9381278f8c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.205715 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(7807384f4402a511d28829969e54fa91f8df902d1affc369202fa9381278f8c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.205755 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(7807384f4402a511d28829969e54fa91f8df902d1affc369202fa9381278f8c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.205800 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-jq2p9_openshift-operators(fc5129c5-6a4c-4d09-807f-1811604f3193)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-jq2p9_openshift-operators(fc5129c5-6a4c-4d09-807f-1811604f3193)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(7807384f4402a511d28829969e54fa91f8df902d1affc369202fa9381278f8c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" podUID="fc5129c5-6a4c-4d09-807f-1811604f3193" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.628724 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" event={"ID":"1aa1f126-12a1-47a0-909b-e2dea899eedd","Type":"ContainerStarted","Data":"157682b82db55e8b2c49e43e543e44738c2e93ed95bf54d83e83deefe1edf9db"} Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.629038 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.629143 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.629351 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.675232 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" podStartSLOduration=7.675211937 podStartE2EDuration="7.675211937s" podCreationTimestamp="2025-11-22 08:13:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:13:10.673183931 +0000 UTC m=+612.277522546" watchObservedRunningTime="2025-11-22 08:13:10.675211937 +0000 UTC m=+612.279550542" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.683808 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.689369 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.902840 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm"] Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.902959 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.903440 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.915678 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz"] Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.915771 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.916221 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.936525 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-jq2p9"] Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.936888 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.937746 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.943970 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5wrtk"] Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.944342 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(0399788d7ed12671426f0b81b27b773a0eb44c9c14aa0023dc6103b96e3010a6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.944391 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.944429 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(0399788d7ed12671426f0b81b27b773a0eb44c9c14aa0023dc6103b96e3010a6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.944482 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(0399788d7ed12671426f0b81b27b773a0eb44c9c14aa0023dc6103b96e3010a6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:10 crc kubenswrapper[4735]: E1122 08:13:10.944545 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators(eab1ae65-df97-4ef5-9876-dccbac6d5304)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators(eab1ae65-df97-4ef5-9876-dccbac6d5304)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(0399788d7ed12671426f0b81b27b773a0eb44c9c14aa0023dc6103b96e3010a6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" podUID="eab1ae65-df97-4ef5-9876-dccbac6d5304" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.945419 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.954748 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579"] Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.954991 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:10 crc kubenswrapper[4735]: I1122 08:13:10.955736 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.021654 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(28d9890c99fe91cc6a1a1893336a162184e4e7c09db02857380490866aa58170): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.021743 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(28d9890c99fe91cc6a1a1893336a162184e4e7c09db02857380490866aa58170): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.021766 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(28d9890c99fe91cc6a1a1893336a162184e4e7c09db02857380490866aa58170): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.021810 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators(73f91b2d-3de6-40da-ac7b-e97c7b63a8ad)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators(73f91b2d-3de6-40da-ac7b-e97c7b63a8ad)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(28d9890c99fe91cc6a1a1893336a162184e4e7c09db02857380490866aa58170): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" podUID="73f91b2d-3de6-40da-ac7b-e97c7b63a8ad" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.026715 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(4d31a0c991692969f117fd43179c71128a38c7a35884d96e792927de409ebbd3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.026775 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(4d31a0c991692969f117fd43179c71128a38c7a35884d96e792927de409ebbd3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.026795 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(4d31a0c991692969f117fd43179c71128a38c7a35884d96e792927de409ebbd3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.026834 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-jq2p9_openshift-operators(fc5129c5-6a4c-4d09-807f-1811604f3193)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-jq2p9_openshift-operators(fc5129c5-6a4c-4d09-807f-1811604f3193)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(4d31a0c991692969f117fd43179c71128a38c7a35884d96e792927de409ebbd3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" podUID="fc5129c5-6a4c-4d09-807f-1811604f3193" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.047663 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(ae167dc665868734f103ae10575482983f7fd09ebb504d78c01fdfe88108c8e2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.047738 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(ae167dc665868734f103ae10575482983f7fd09ebb504d78c01fdfe88108c8e2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.047758 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(ae167dc665868734f103ae10575482983f7fd09ebb504d78c01fdfe88108c8e2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.047794 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-5wrtk_openshift-operators(93273ee2-9156-4b65-abe9-1e020aa4ea55)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-5wrtk_openshift-operators(93273ee2-9156-4b65-abe9-1e020aa4ea55)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(ae167dc665868734f103ae10575482983f7fd09ebb504d78c01fdfe88108c8e2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" podUID="93273ee2-9156-4b65-abe9-1e020aa4ea55" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.066205 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(80ed225ad604f823fe6a6222a94a524dc2b8cbc2ef5f46efaecb7129f9341c33): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.066276 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(80ed225ad604f823fe6a6222a94a524dc2b8cbc2ef5f46efaecb7129f9341c33): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.066300 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(80ed225ad604f823fe6a6222a94a524dc2b8cbc2ef5f46efaecb7129f9341c33): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:11 crc kubenswrapper[4735]: E1122 08:13:11.066339 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators(50c1fd51-0497-44e6-b9d6-52b48e5a4133)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators(50c1fd51-0497-44e6-b9d6-52b48e5a4133)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(80ed225ad604f823fe6a6222a94a524dc2b8cbc2ef5f46efaecb7129f9341c33): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" podUID="50c1fd51-0497-44e6-b9d6-52b48e5a4133" Nov 22 08:13:13 crc kubenswrapper[4735]: I1122 08:13:13.263551 4735 scope.go:117] "RemoveContainer" containerID="8f71d0c8bea2f1c5a42c8365198e3fe40631e28bfc7c5f174a032e124a354f1a" Nov 22 08:13:13 crc kubenswrapper[4735]: E1122 08:13:13.264081 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-85z2h_openshift-multus(61dbab7f-c903-4b55-9d90-aacfd12f7f0f)\"" pod="openshift-multus/multus-85z2h" podUID="61dbab7f-c903-4b55-9d90-aacfd12f7f0f" Nov 22 08:13:22 crc kubenswrapper[4735]: I1122 08:13:22.263177 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:22 crc kubenswrapper[4735]: I1122 08:13:22.263323 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:22 crc kubenswrapper[4735]: I1122 08:13:22.265042 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:22 crc kubenswrapper[4735]: I1122 08:13:22.265189 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:22 crc kubenswrapper[4735]: E1122 08:13:22.302420 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(e266f7801656dc89f5f6ae97ec52bd99214596c04238176779510e3c914f4fd2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:22 crc kubenswrapper[4735]: E1122 08:13:22.302513 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(e266f7801656dc89f5f6ae97ec52bd99214596c04238176779510e3c914f4fd2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:22 crc kubenswrapper[4735]: E1122 08:13:22.302543 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(e266f7801656dc89f5f6ae97ec52bd99214596c04238176779510e3c914f4fd2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:22 crc kubenswrapper[4735]: E1122 08:13:22.302608 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators(eab1ae65-df97-4ef5-9876-dccbac6d5304)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators(eab1ae65-df97-4ef5-9876-dccbac6d5304)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-sxnpm_openshift-operators_eab1ae65-df97-4ef5-9876-dccbac6d5304_0(e266f7801656dc89f5f6ae97ec52bd99214596c04238176779510e3c914f4fd2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" podUID="eab1ae65-df97-4ef5-9876-dccbac6d5304" Nov 22 08:13:22 crc kubenswrapper[4735]: E1122 08:13:22.310357 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(c07ff825b4187f24184a7898eb38453183bc667dee8671790b757e51dbe77dde): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:22 crc kubenswrapper[4735]: E1122 08:13:22.310418 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(c07ff825b4187f24184a7898eb38453183bc667dee8671790b757e51dbe77dde): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:22 crc kubenswrapper[4735]: E1122 08:13:22.310437 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(c07ff825b4187f24184a7898eb38453183bc667dee8671790b757e51dbe77dde): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:22 crc kubenswrapper[4735]: E1122 08:13:22.310499 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-5wrtk_openshift-operators(93273ee2-9156-4b65-abe9-1e020aa4ea55)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-5wrtk_openshift-operators(93273ee2-9156-4b65-abe9-1e020aa4ea55)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5wrtk_openshift-operators_93273ee2-9156-4b65-abe9-1e020aa4ea55_0(c07ff825b4187f24184a7898eb38453183bc667dee8671790b757e51dbe77dde): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" podUID="93273ee2-9156-4b65-abe9-1e020aa4ea55" Nov 22 08:13:23 crc kubenswrapper[4735]: I1122 08:13:23.262906 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:23 crc kubenswrapper[4735]: I1122 08:13:23.263720 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:23 crc kubenswrapper[4735]: E1122 08:13:23.293208 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(77cec27903fe9ed7e4de6b32fca8bdc8e1e173fa2ff3a10489180bce122b1784): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:23 crc kubenswrapper[4735]: E1122 08:13:23.293290 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(77cec27903fe9ed7e4de6b32fca8bdc8e1e173fa2ff3a10489180bce122b1784): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:23 crc kubenswrapper[4735]: E1122 08:13:23.293318 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(77cec27903fe9ed7e4de6b32fca8bdc8e1e173fa2ff3a10489180bce122b1784): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:23 crc kubenswrapper[4735]: E1122 08:13:23.293372 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-jq2p9_openshift-operators(fc5129c5-6a4c-4d09-807f-1811604f3193)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-jq2p9_openshift-operators(fc5129c5-6a4c-4d09-807f-1811604f3193)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-jq2p9_openshift-operators_fc5129c5-6a4c-4d09-807f-1811604f3193_0(77cec27903fe9ed7e4de6b32fca8bdc8e1e173fa2ff3a10489180bce122b1784): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" podUID="fc5129c5-6a4c-4d09-807f-1811604f3193" Nov 22 08:13:24 crc kubenswrapper[4735]: I1122 08:13:24.263097 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:24 crc kubenswrapper[4735]: I1122 08:13:24.263537 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:24 crc kubenswrapper[4735]: E1122 08:13:24.287551 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(ee53866f717fa0a4b021e1b036d7705a7d8cc752e50f3fb8e16d1aa5a2fcfed1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:24 crc kubenswrapper[4735]: E1122 08:13:24.288279 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(ee53866f717fa0a4b021e1b036d7705a7d8cc752e50f3fb8e16d1aa5a2fcfed1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:24 crc kubenswrapper[4735]: E1122 08:13:24.288393 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(ee53866f717fa0a4b021e1b036d7705a7d8cc752e50f3fb8e16d1aa5a2fcfed1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:24 crc kubenswrapper[4735]: E1122 08:13:24.288551 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators(50c1fd51-0497-44e6-b9d6-52b48e5a4133)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators(50c1fd51-0497-44e6-b9d6-52b48e5a4133)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_openshift-operators_50c1fd51-0497-44e6-b9d6-52b48e5a4133_0(ee53866f717fa0a4b021e1b036d7705a7d8cc752e50f3fb8e16d1aa5a2fcfed1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" podUID="50c1fd51-0497-44e6-b9d6-52b48e5a4133" Nov 22 08:13:25 crc kubenswrapper[4735]: I1122 08:13:25.262922 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:25 crc kubenswrapper[4735]: I1122 08:13:25.263825 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:25 crc kubenswrapper[4735]: E1122 08:13:25.290969 4735 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(712e9416cfa061cffd70ec761516a5a9a978763c0396fdec4ef0031745a70f3a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 08:13:25 crc kubenswrapper[4735]: E1122 08:13:25.291035 4735 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(712e9416cfa061cffd70ec761516a5a9a978763c0396fdec4ef0031745a70f3a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:25 crc kubenswrapper[4735]: E1122 08:13:25.291057 4735 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(712e9416cfa061cffd70ec761516a5a9a978763c0396fdec4ef0031745a70f3a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:25 crc kubenswrapper[4735]: E1122 08:13:25.291100 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators(73f91b2d-3de6-40da-ac7b-e97c7b63a8ad)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators(73f91b2d-3de6-40da-ac7b-e97c7b63a8ad)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_openshift-operators_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad_0(712e9416cfa061cffd70ec761516a5a9a978763c0396fdec4ef0031745a70f3a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" podUID="73f91b2d-3de6-40da-ac7b-e97c7b63a8ad" Nov 22 08:13:26 crc kubenswrapper[4735]: I1122 08:13:26.263719 4735 scope.go:117] "RemoveContainer" containerID="8f71d0c8bea2f1c5a42c8365198e3fe40631e28bfc7c5f174a032e124a354f1a" Nov 22 08:13:26 crc kubenswrapper[4735]: I1122 08:13:26.726906 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-85z2h_61dbab7f-c903-4b55-9d90-aacfd12f7f0f/kube-multus/2.log" Nov 22 08:13:26 crc kubenswrapper[4735]: I1122 08:13:26.727202 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-85z2h" event={"ID":"61dbab7f-c903-4b55-9d90-aacfd12f7f0f","Type":"ContainerStarted","Data":"407dd060b290e045f6642cd928fbf52d820c50346b2d1c13fce8d63330248cc8"} Nov 22 08:13:33 crc kubenswrapper[4735]: I1122 08:13:33.263597 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:33 crc kubenswrapper[4735]: I1122 08:13:33.263593 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:33 crc kubenswrapper[4735]: I1122 08:13:33.264424 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" Nov 22 08:13:33 crc kubenswrapper[4735]: I1122 08:13:33.264893 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:33 crc kubenswrapper[4735]: I1122 08:13:33.657859 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5wrtk"] Nov 22 08:13:33 crc kubenswrapper[4735]: I1122 08:13:33.766269 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" event={"ID":"93273ee2-9156-4b65-abe9-1e020aa4ea55","Type":"ContainerStarted","Data":"c863f299fa671e3335d27d1ebb583ffc90cc5a0b9f94fac7e0a648281ce1511b"} Nov 22 08:13:33 crc kubenswrapper[4735]: I1122 08:13:33.791315 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm"] Nov 22 08:13:33 crc kubenswrapper[4735]: W1122 08:13:33.795552 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeab1ae65_df97_4ef5_9876_dccbac6d5304.slice/crio-ca94889fefc0cf44189c054ad9b9bf313ec72fe7b314e648eb66204e6ec12b24 WatchSource:0}: Error finding container ca94889fefc0cf44189c054ad9b9bf313ec72fe7b314e648eb66204e6ec12b24: Status 404 returned error can't find the container with id ca94889fefc0cf44189c054ad9b9bf313ec72fe7b314e648eb66204e6ec12b24 Nov 22 08:13:33 crc kubenswrapper[4735]: I1122 08:13:33.861856 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4ppq" Nov 22 08:13:34 crc kubenswrapper[4735]: I1122 08:13:34.774618 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" event={"ID":"eab1ae65-df97-4ef5-9876-dccbac6d5304","Type":"ContainerStarted","Data":"ca94889fefc0cf44189c054ad9b9bf313ec72fe7b314e648eb66204e6ec12b24"} Nov 22 08:13:37 crc kubenswrapper[4735]: I1122 08:13:37.264209 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:37 crc kubenswrapper[4735]: I1122 08:13:37.264826 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:37 crc kubenswrapper[4735]: I1122 08:13:37.265001 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:37 crc kubenswrapper[4735]: I1122 08:13:37.265325 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" Nov 22 08:13:40 crc kubenswrapper[4735]: I1122 08:13:40.262916 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:40 crc kubenswrapper[4735]: I1122 08:13:40.263402 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.729832 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-jq2p9"] Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.742098 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz"] Nov 22 08:13:42 crc kubenswrapper[4735]: W1122 08:13:42.755853 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73f91b2d_3de6_40da_ac7b_e97c7b63a8ad.slice/crio-f72abb3fffc5e78a527df06a347d00d2f2d784dd8575a06616314e6537b11286 WatchSource:0}: Error finding container f72abb3fffc5e78a527df06a347d00d2f2d784dd8575a06616314e6537b11286: Status 404 returned error can't find the container with id f72abb3fffc5e78a527df06a347d00d2f2d784dd8575a06616314e6537b11286 Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.810440 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579"] Nov 22 08:13:42 crc kubenswrapper[4735]: W1122 08:13:42.814734 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50c1fd51_0497_44e6_b9d6_52b48e5a4133.slice/crio-b1fe4093244443b8eadc49dcce3fed69230a9150dbf2445a39b1a822355a6a5f WatchSource:0}: Error finding container b1fe4093244443b8eadc49dcce3fed69230a9150dbf2445a39b1a822355a6a5f: Status 404 returned error can't find the container with id b1fe4093244443b8eadc49dcce3fed69230a9150dbf2445a39b1a822355a6a5f Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.843185 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" event={"ID":"93273ee2-9156-4b65-abe9-1e020aa4ea55","Type":"ContainerStarted","Data":"d28479815f64b3f95c0803de3b4cd1345c758fc2e6d2e6cb4c907179c5362926"} Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.843596 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.844364 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" event={"ID":"50c1fd51-0497-44e6-b9d6-52b48e5a4133","Type":"ContainerStarted","Data":"b1fe4093244443b8eadc49dcce3fed69230a9150dbf2445a39b1a822355a6a5f"} Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.848063 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" event={"ID":"73f91b2d-3de6-40da-ac7b-e97c7b63a8ad","Type":"ContainerStarted","Data":"f72abb3fffc5e78a527df06a347d00d2f2d784dd8575a06616314e6537b11286"} Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.849482 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" event={"ID":"eab1ae65-df97-4ef5-9876-dccbac6d5304","Type":"ContainerStarted","Data":"b2ccb85e4e898a7d88853cca83d2fb2d3d5fbdf91beff6330cb1b707358e963e"} Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.850411 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" event={"ID":"fc5129c5-6a4c-4d09-807f-1811604f3193","Type":"ContainerStarted","Data":"1de5136b4dfb2ae933a4506fda616ca338dcea09851069e3a47f75f7430bf174"} Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.853706 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.890262 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-sxnpm" podStartSLOduration=25.401618509 podStartE2EDuration="33.890243011s" podCreationTimestamp="2025-11-22 08:13:09 +0000 UTC" firstStartedPulling="2025-11-22 08:13:33.79808598 +0000 UTC m=+635.402424585" lastFinishedPulling="2025-11-22 08:13:42.286710482 +0000 UTC m=+643.891049087" observedRunningTime="2025-11-22 08:13:42.887249309 +0000 UTC m=+644.491587924" watchObservedRunningTime="2025-11-22 08:13:42.890243011 +0000 UTC m=+644.494581616" Nov 22 08:13:42 crc kubenswrapper[4735]: I1122 08:13:42.892572 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-5wrtk" podStartSLOduration=25.244385145 podStartE2EDuration="33.892562593s" podCreationTimestamp="2025-11-22 08:13:09 +0000 UTC" firstStartedPulling="2025-11-22 08:13:33.672698056 +0000 UTC m=+635.277036661" lastFinishedPulling="2025-11-22 08:13:42.320875504 +0000 UTC m=+643.925214109" observedRunningTime="2025-11-22 08:13:42.864143458 +0000 UTC m=+644.468482063" watchObservedRunningTime="2025-11-22 08:13:42.892562593 +0000 UTC m=+644.496901198" Nov 22 08:13:45 crc kubenswrapper[4735]: I1122 08:13:45.871879 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" event={"ID":"73f91b2d-3de6-40da-ac7b-e97c7b63a8ad","Type":"ContainerStarted","Data":"769357b1025518fd0116faa77b9366d0354e05abc094eea772c746fab712011c"} Nov 22 08:13:45 crc kubenswrapper[4735]: I1122 08:13:45.873941 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" event={"ID":"fc5129c5-6a4c-4d09-807f-1811604f3193","Type":"ContainerStarted","Data":"7f4455428b6f80349afb913e36632e920fb8c00d488350407931f7077557677e"} Nov 22 08:13:45 crc kubenswrapper[4735]: I1122 08:13:45.874306 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:45 crc kubenswrapper[4735]: I1122 08:13:45.875908 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" event={"ID":"50c1fd51-0497-44e6-b9d6-52b48e5a4133","Type":"ContainerStarted","Data":"7263d8309e79b404f8833403bca587562fed6b3ef5684511a6e82a2c3da80d38"} Nov 22 08:13:45 crc kubenswrapper[4735]: I1122 08:13:45.898836 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz" podStartSLOduration=34.620478279 podStartE2EDuration="36.898817527s" podCreationTimestamp="2025-11-22 08:13:09 +0000 UTC" firstStartedPulling="2025-11-22 08:13:42.759155491 +0000 UTC m=+644.363494096" lastFinishedPulling="2025-11-22 08:13:45.037494739 +0000 UTC m=+646.641833344" observedRunningTime="2025-11-22 08:13:45.896143213 +0000 UTC m=+647.500481858" watchObservedRunningTime="2025-11-22 08:13:45.898817527 +0000 UTC m=+647.503156122" Nov 22 08:13:45 crc kubenswrapper[4735]: I1122 08:13:45.916748 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" podStartSLOduration=34.605842428 podStartE2EDuration="36.916731605s" podCreationTimestamp="2025-11-22 08:13:09 +0000 UTC" firstStartedPulling="2025-11-22 08:13:42.739855024 +0000 UTC m=+644.344193629" lastFinishedPulling="2025-11-22 08:13:45.050744201 +0000 UTC m=+646.655082806" observedRunningTime="2025-11-22 08:13:45.911012639 +0000 UTC m=+647.515351244" watchObservedRunningTime="2025-11-22 08:13:45.916731605 +0000 UTC m=+647.521070210" Nov 22 08:13:45 crc kubenswrapper[4735]: I1122 08:13:45.947056 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7d649cf67c-kj579" podStartSLOduration=34.720265673 podStartE2EDuration="36.947034583s" podCreationTimestamp="2025-11-22 08:13:09 +0000 UTC" firstStartedPulling="2025-11-22 08:13:42.81700929 +0000 UTC m=+644.421347895" lastFinishedPulling="2025-11-22 08:13:45.0437782 +0000 UTC m=+646.648116805" observedRunningTime="2025-11-22 08:13:45.943746733 +0000 UTC m=+647.548085378" watchObservedRunningTime="2025-11-22 08:13:45.947034583 +0000 UTC m=+647.551373198" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.180769 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-jq2p9" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.818113 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-t6b7j"] Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.819132 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-t6b7j" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.821187 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.821202 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.826153 4735 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-nzv5n" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.826164 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-t6b7j"] Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.828214 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dbnm5"] Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.829052 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-dbnm5" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.832133 4735 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-znskw" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.849039 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dbnm5"] Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.854938 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-26wft"] Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.856317 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.863700 4735 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-94455" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.871490 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-26wft"] Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.995165 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6j6k\" (UniqueName: \"kubernetes.io/projected/24a36a64-01ec-4f38-b118-e1c6097e8965-kube-api-access-j6j6k\") pod \"cert-manager-webhook-5655c58dd6-26wft\" (UID: \"24a36a64-01ec-4f38-b118-e1c6097e8965\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.995256 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tfmj\" (UniqueName: \"kubernetes.io/projected/cccf1779-d468-4e20-83ee-6a7c85b1d7ab-kube-api-access-7tfmj\") pod \"cert-manager-5b446d88c5-dbnm5\" (UID: \"cccf1779-d468-4e20-83ee-6a7c85b1d7ab\") " pod="cert-manager/cert-manager-5b446d88c5-dbnm5" Nov 22 08:13:50 crc kubenswrapper[4735]: I1122 08:13:50.995309 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87ssj\" (UniqueName: \"kubernetes.io/projected/9a904776-a627-4df0-abf1-1832ba4a064c-kube-api-access-87ssj\") pod \"cert-manager-cainjector-7f985d654d-t6b7j\" (UID: \"9a904776-a627-4df0-abf1-1832ba4a064c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-t6b7j" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.096597 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87ssj\" (UniqueName: \"kubernetes.io/projected/9a904776-a627-4df0-abf1-1832ba4a064c-kube-api-access-87ssj\") pod \"cert-manager-cainjector-7f985d654d-t6b7j\" (UID: \"9a904776-a627-4df0-abf1-1832ba4a064c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-t6b7j" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.096717 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6j6k\" (UniqueName: \"kubernetes.io/projected/24a36a64-01ec-4f38-b118-e1c6097e8965-kube-api-access-j6j6k\") pod \"cert-manager-webhook-5655c58dd6-26wft\" (UID: \"24a36a64-01ec-4f38-b118-e1c6097e8965\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.096780 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tfmj\" (UniqueName: \"kubernetes.io/projected/cccf1779-d468-4e20-83ee-6a7c85b1d7ab-kube-api-access-7tfmj\") pod \"cert-manager-5b446d88c5-dbnm5\" (UID: \"cccf1779-d468-4e20-83ee-6a7c85b1d7ab\") " pod="cert-manager/cert-manager-5b446d88c5-dbnm5" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.116425 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6j6k\" (UniqueName: \"kubernetes.io/projected/24a36a64-01ec-4f38-b118-e1c6097e8965-kube-api-access-j6j6k\") pod \"cert-manager-webhook-5655c58dd6-26wft\" (UID: \"24a36a64-01ec-4f38-b118-e1c6097e8965\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.116832 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87ssj\" (UniqueName: \"kubernetes.io/projected/9a904776-a627-4df0-abf1-1832ba4a064c-kube-api-access-87ssj\") pod \"cert-manager-cainjector-7f985d654d-t6b7j\" (UID: \"9a904776-a627-4df0-abf1-1832ba4a064c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-t6b7j" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.117111 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tfmj\" (UniqueName: \"kubernetes.io/projected/cccf1779-d468-4e20-83ee-6a7c85b1d7ab-kube-api-access-7tfmj\") pod \"cert-manager-5b446d88c5-dbnm5\" (UID: \"cccf1779-d468-4e20-83ee-6a7c85b1d7ab\") " pod="cert-manager/cert-manager-5b446d88c5-dbnm5" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.141156 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-t6b7j" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.153514 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-dbnm5" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.175505 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.532587 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-t6b7j"] Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.615398 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-26wft"] Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.622275 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dbnm5"] Nov 22 08:13:51 crc kubenswrapper[4735]: W1122 08:13:51.648356 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24a36a64_01ec_4f38_b118_e1c6097e8965.slice/crio-7db7d0a6384db0ba41802f993f7841f635592b9b128322fb2c3d43f9603c07b6 WatchSource:0}: Error finding container 7db7d0a6384db0ba41802f993f7841f635592b9b128322fb2c3d43f9603c07b6: Status 404 returned error can't find the container with id 7db7d0a6384db0ba41802f993f7841f635592b9b128322fb2c3d43f9603c07b6 Nov 22 08:13:51 crc kubenswrapper[4735]: W1122 08:13:51.649407 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcccf1779_d468_4e20_83ee_6a7c85b1d7ab.slice/crio-ce4da42150dfd88f97760ae501b7997629a0d5d060d4ccce04fd68351d4eb715 WatchSource:0}: Error finding container ce4da42150dfd88f97760ae501b7997629a0d5d060d4ccce04fd68351d4eb715: Status 404 returned error can't find the container with id ce4da42150dfd88f97760ae501b7997629a0d5d060d4ccce04fd68351d4eb715 Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.911693 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" event={"ID":"24a36a64-01ec-4f38-b118-e1c6097e8965","Type":"ContainerStarted","Data":"7db7d0a6384db0ba41802f993f7841f635592b9b128322fb2c3d43f9603c07b6"} Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.913031 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-dbnm5" event={"ID":"cccf1779-d468-4e20-83ee-6a7c85b1d7ab","Type":"ContainerStarted","Data":"ce4da42150dfd88f97760ae501b7997629a0d5d060d4ccce04fd68351d4eb715"} Nov 22 08:13:51 crc kubenswrapper[4735]: I1122 08:13:51.914227 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-t6b7j" event={"ID":"9a904776-a627-4df0-abf1-1832ba4a064c","Type":"ContainerStarted","Data":"5319aac5afad184f581d8fd4db437d6400c73f01de6dc31e298af989c1909858"} Nov 22 08:13:56 crc kubenswrapper[4735]: I1122 08:13:56.957322 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-dbnm5" event={"ID":"cccf1779-d468-4e20-83ee-6a7c85b1d7ab","Type":"ContainerStarted","Data":"3ad6cb4100f9e671854be7de601a337636ef4f775731606af9d1a842c17ffe0f"} Nov 22 08:13:56 crc kubenswrapper[4735]: I1122 08:13:56.961510 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-t6b7j" event={"ID":"9a904776-a627-4df0-abf1-1832ba4a064c","Type":"ContainerStarted","Data":"4b79efbeda45a69039039323634390efd0a00aa07ad9f7a72f45af93d5f9d086"} Nov 22 08:13:56 crc kubenswrapper[4735]: I1122 08:13:56.963662 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" event={"ID":"24a36a64-01ec-4f38-b118-e1c6097e8965","Type":"ContainerStarted","Data":"468fc9d6b5dde431b6c92461fd4a761630fab764283d40207874d21e53971347"} Nov 22 08:13:56 crc kubenswrapper[4735]: I1122 08:13:56.963857 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" Nov 22 08:13:56 crc kubenswrapper[4735]: I1122 08:13:56.980279 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-dbnm5" podStartSLOduration=2.327644795 podStartE2EDuration="6.980261624s" podCreationTimestamp="2025-11-22 08:13:50 +0000 UTC" firstStartedPulling="2025-11-22 08:13:51.651122744 +0000 UTC m=+653.255461349" lastFinishedPulling="2025-11-22 08:13:56.303739573 +0000 UTC m=+657.908078178" observedRunningTime="2025-11-22 08:13:56.977547749 +0000 UTC m=+658.581886374" watchObservedRunningTime="2025-11-22 08:13:56.980261624 +0000 UTC m=+658.584600229" Nov 22 08:13:57 crc kubenswrapper[4735]: I1122 08:13:57.006863 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-t6b7j" podStartSLOduration=2.257400821 podStartE2EDuration="7.00684445s" podCreationTimestamp="2025-11-22 08:13:50 +0000 UTC" firstStartedPulling="2025-11-22 08:13:51.547901036 +0000 UTC m=+653.152239641" lastFinishedPulling="2025-11-22 08:13:56.297344665 +0000 UTC m=+657.901683270" observedRunningTime="2025-11-22 08:13:57.004605778 +0000 UTC m=+658.608944403" watchObservedRunningTime="2025-11-22 08:13:57.00684445 +0000 UTC m=+658.611183055" Nov 22 08:13:57 crc kubenswrapper[4735]: I1122 08:13:57.033253 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" podStartSLOduration=2.319440937 podStartE2EDuration="7.03322595s" podCreationTimestamp="2025-11-22 08:13:50 +0000 UTC" firstStartedPulling="2025-11-22 08:13:51.651095333 +0000 UTC m=+653.255433938" lastFinishedPulling="2025-11-22 08:13:56.364880346 +0000 UTC m=+657.969218951" observedRunningTime="2025-11-22 08:13:57.032630764 +0000 UTC m=+658.636969369" watchObservedRunningTime="2025-11-22 08:13:57.03322595 +0000 UTC m=+658.637564575" Nov 22 08:14:01 crc kubenswrapper[4735]: I1122 08:14:01.178494 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.328652 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k"] Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.330450 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.332370 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.341733 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k"] Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.434766 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c24vq\" (UniqueName: \"kubernetes.io/projected/73825640-996d-43e1-8e5d-e5ed2a514896-kube-api-access-c24vq\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.434905 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.434972 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.530692 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6"] Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.532310 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.535832 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c24vq\" (UniqueName: \"kubernetes.io/projected/73825640-996d-43e1-8e5d-e5ed2a514896-kube-api-access-c24vq\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.535893 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlfq2\" (UniqueName: \"kubernetes.io/projected/467ea8b5-523e-40a4-8cd2-9f20f203a975-kube-api-access-nlfq2\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.535931 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.535977 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.536042 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.536071 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.536393 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.536508 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.541655 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6"] Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.558814 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c24vq\" (UniqueName: \"kubernetes.io/projected/73825640-996d-43e1-8e5d-e5ed2a514896-kube-api-access-c24vq\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.637260 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.637623 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.637709 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlfq2\" (UniqueName: \"kubernetes.io/projected/467ea8b5-523e-40a4-8cd2-9f20f203a975-kube-api-access-nlfq2\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.637950 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.638236 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.658215 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.665072 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlfq2\" (UniqueName: \"kubernetes.io/projected/467ea8b5-523e-40a4-8cd2-9f20f203a975-kube-api-access-nlfq2\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.848550 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:28 crc kubenswrapper[4735]: I1122 08:14:28.878936 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k"] Nov 22 08:14:29 crc kubenswrapper[4735]: I1122 08:14:29.133800 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6"] Nov 22 08:14:29 crc kubenswrapper[4735]: I1122 08:14:29.177520 4735 generic.go:334] "Generic (PLEG): container finished" podID="73825640-996d-43e1-8e5d-e5ed2a514896" containerID="db7824d358300c58b9ce729217f199414832cac5bfda599c8f37b503803d8482" exitCode=0 Nov 22 08:14:29 crc kubenswrapper[4735]: I1122 08:14:29.177677 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" event={"ID":"73825640-996d-43e1-8e5d-e5ed2a514896","Type":"ContainerDied","Data":"db7824d358300c58b9ce729217f199414832cac5bfda599c8f37b503803d8482"} Nov 22 08:14:29 crc kubenswrapper[4735]: I1122 08:14:29.177796 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" event={"ID":"73825640-996d-43e1-8e5d-e5ed2a514896","Type":"ContainerStarted","Data":"d6f7719b7fe05134cfb24dba4dfd02ad43af3943ee21e1116ff337db63b06033"} Nov 22 08:14:29 crc kubenswrapper[4735]: W1122 08:14:29.219200 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod467ea8b5_523e_40a4_8cd2_9f20f203a975.slice/crio-5e98b7cc18a1228171d2009db658e86bd71e7b99b98b896f3db35a66cadeac3c WatchSource:0}: Error finding container 5e98b7cc18a1228171d2009db658e86bd71e7b99b98b896f3db35a66cadeac3c: Status 404 returned error can't find the container with id 5e98b7cc18a1228171d2009db658e86bd71e7b99b98b896f3db35a66cadeac3c Nov 22 08:14:30 crc kubenswrapper[4735]: I1122 08:14:30.184345 4735 generic.go:334] "Generic (PLEG): container finished" podID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerID="080976996e88a309cb80091d861f5cffac6ddc8aeda219015d43263af6c8fc74" exitCode=0 Nov 22 08:14:30 crc kubenswrapper[4735]: I1122 08:14:30.184523 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" event={"ID":"467ea8b5-523e-40a4-8cd2-9f20f203a975","Type":"ContainerDied","Data":"080976996e88a309cb80091d861f5cffac6ddc8aeda219015d43263af6c8fc74"} Nov 22 08:14:30 crc kubenswrapper[4735]: I1122 08:14:30.184646 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" event={"ID":"467ea8b5-523e-40a4-8cd2-9f20f203a975","Type":"ContainerStarted","Data":"5e98b7cc18a1228171d2009db658e86bd71e7b99b98b896f3db35a66cadeac3c"} Nov 22 08:14:31 crc kubenswrapper[4735]: I1122 08:14:31.200967 4735 generic.go:334] "Generic (PLEG): container finished" podID="73825640-996d-43e1-8e5d-e5ed2a514896" containerID="313e6f97d7926f952cd53c35b16f3988bfadc7c18c16bf1e4292a8088bdd77cc" exitCode=0 Nov 22 08:14:31 crc kubenswrapper[4735]: I1122 08:14:31.201011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" event={"ID":"73825640-996d-43e1-8e5d-e5ed2a514896","Type":"ContainerDied","Data":"313e6f97d7926f952cd53c35b16f3988bfadc7c18c16bf1e4292a8088bdd77cc"} Nov 22 08:14:32 crc kubenswrapper[4735]: I1122 08:14:32.208121 4735 generic.go:334] "Generic (PLEG): container finished" podID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerID="f2514a3a8fbe801e661220bfe5c055fae2f9ecd35159a51b351031609120c14e" exitCode=0 Nov 22 08:14:32 crc kubenswrapper[4735]: I1122 08:14:32.208180 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" event={"ID":"467ea8b5-523e-40a4-8cd2-9f20f203a975","Type":"ContainerDied","Data":"f2514a3a8fbe801e661220bfe5c055fae2f9ecd35159a51b351031609120c14e"} Nov 22 08:14:32 crc kubenswrapper[4735]: I1122 08:14:32.211512 4735 generic.go:334] "Generic (PLEG): container finished" podID="73825640-996d-43e1-8e5d-e5ed2a514896" containerID="af324db404ace3646738780b9c79530e174114641618cf42c259907468b13032" exitCode=0 Nov 22 08:14:32 crc kubenswrapper[4735]: I1122 08:14:32.211555 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" event={"ID":"73825640-996d-43e1-8e5d-e5ed2a514896","Type":"ContainerDied","Data":"af324db404ace3646738780b9c79530e174114641618cf42c259907468b13032"} Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.231753 4735 generic.go:334] "Generic (PLEG): container finished" podID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerID="7463a4922e7d6ea4670bdbb3e41cf9078bd9f79c13d82106811a43ee25cd8ada" exitCode=0 Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.231923 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" event={"ID":"467ea8b5-523e-40a4-8cd2-9f20f203a975","Type":"ContainerDied","Data":"7463a4922e7d6ea4670bdbb3e41cf9078bd9f79c13d82106811a43ee25cd8ada"} Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.536028 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.714615 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-bundle\") pod \"73825640-996d-43e1-8e5d-e5ed2a514896\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.714707 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c24vq\" (UniqueName: \"kubernetes.io/projected/73825640-996d-43e1-8e5d-e5ed2a514896-kube-api-access-c24vq\") pod \"73825640-996d-43e1-8e5d-e5ed2a514896\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.714779 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-util\") pod \"73825640-996d-43e1-8e5d-e5ed2a514896\" (UID: \"73825640-996d-43e1-8e5d-e5ed2a514896\") " Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.715493 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-bundle" (OuterVolumeSpecName: "bundle") pod "73825640-996d-43e1-8e5d-e5ed2a514896" (UID: "73825640-996d-43e1-8e5d-e5ed2a514896"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.721715 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73825640-996d-43e1-8e5d-e5ed2a514896-kube-api-access-c24vq" (OuterVolumeSpecName: "kube-api-access-c24vq") pod "73825640-996d-43e1-8e5d-e5ed2a514896" (UID: "73825640-996d-43e1-8e5d-e5ed2a514896"). InnerVolumeSpecName "kube-api-access-c24vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.734520 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-util" (OuterVolumeSpecName: "util") pod "73825640-996d-43e1-8e5d-e5ed2a514896" (UID: "73825640-996d-43e1-8e5d-e5ed2a514896"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.816057 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.816094 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c24vq\" (UniqueName: \"kubernetes.io/projected/73825640-996d-43e1-8e5d-e5ed2a514896-kube-api-access-c24vq\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:33 crc kubenswrapper[4735]: I1122 08:14:33.816111 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73825640-996d-43e1-8e5d-e5ed2a514896-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.241653 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.241648 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k" event={"ID":"73825640-996d-43e1-8e5d-e5ed2a514896","Type":"ContainerDied","Data":"d6f7719b7fe05134cfb24dba4dfd02ad43af3943ee21e1116ff337db63b06033"} Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.242111 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6f7719b7fe05134cfb24dba4dfd02ad43af3943ee21e1116ff337db63b06033" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.476203 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.624965 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-bundle\") pod \"467ea8b5-523e-40a4-8cd2-9f20f203a975\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.625120 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlfq2\" (UniqueName: \"kubernetes.io/projected/467ea8b5-523e-40a4-8cd2-9f20f203a975-kube-api-access-nlfq2\") pod \"467ea8b5-523e-40a4-8cd2-9f20f203a975\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.625163 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-util\") pod \"467ea8b5-523e-40a4-8cd2-9f20f203a975\" (UID: \"467ea8b5-523e-40a4-8cd2-9f20f203a975\") " Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.625889 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-bundle" (OuterVolumeSpecName: "bundle") pod "467ea8b5-523e-40a4-8cd2-9f20f203a975" (UID: "467ea8b5-523e-40a4-8cd2-9f20f203a975"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.631504 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/467ea8b5-523e-40a4-8cd2-9f20f203a975-kube-api-access-nlfq2" (OuterVolumeSpecName: "kube-api-access-nlfq2") pod "467ea8b5-523e-40a4-8cd2-9f20f203a975" (UID: "467ea8b5-523e-40a4-8cd2-9f20f203a975"). InnerVolumeSpecName "kube-api-access-nlfq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.638838 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-util" (OuterVolumeSpecName: "util") pod "467ea8b5-523e-40a4-8cd2-9f20f203a975" (UID: "467ea8b5-523e-40a4-8cd2-9f20f203a975"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.726483 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.726726 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlfq2\" (UniqueName: \"kubernetes.io/projected/467ea8b5-523e-40a4-8cd2-9f20f203a975-kube-api-access-nlfq2\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:34 crc kubenswrapper[4735]: I1122 08:14:34.726805 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/467ea8b5-523e-40a4-8cd2-9f20f203a975-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:35 crc kubenswrapper[4735]: I1122 08:14:35.249846 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" event={"ID":"467ea8b5-523e-40a4-8cd2-9f20f203a975","Type":"ContainerDied","Data":"5e98b7cc18a1228171d2009db658e86bd71e7b99b98b896f3db35a66cadeac3c"} Nov 22 08:14:35 crc kubenswrapper[4735]: I1122 08:14:35.250677 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e98b7cc18a1228171d2009db658e86bd71e7b99b98b896f3db35a66cadeac3c" Nov 22 08:14:35 crc kubenswrapper[4735]: I1122 08:14:35.249902 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540024 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-cp7lm"] Nov 22 08:14:38 crc kubenswrapper[4735]: E1122 08:14:38.540555 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73825640-996d-43e1-8e5d-e5ed2a514896" containerName="extract" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540566 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="73825640-996d-43e1-8e5d-e5ed2a514896" containerName="extract" Nov 22 08:14:38 crc kubenswrapper[4735]: E1122 08:14:38.540579 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73825640-996d-43e1-8e5d-e5ed2a514896" containerName="pull" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540585 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="73825640-996d-43e1-8e5d-e5ed2a514896" containerName="pull" Nov 22 08:14:38 crc kubenswrapper[4735]: E1122 08:14:38.540596 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerName="util" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540602 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerName="util" Nov 22 08:14:38 crc kubenswrapper[4735]: E1122 08:14:38.540616 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73825640-996d-43e1-8e5d-e5ed2a514896" containerName="util" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540622 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="73825640-996d-43e1-8e5d-e5ed2a514896" containerName="util" Nov 22 08:14:38 crc kubenswrapper[4735]: E1122 08:14:38.540630 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerName="extract" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540636 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerName="extract" Nov 22 08:14:38 crc kubenswrapper[4735]: E1122 08:14:38.540645 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerName="pull" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540651 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerName="pull" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540770 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="467ea8b5-523e-40a4-8cd2-9f20f203a975" containerName="extract" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.540782 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="73825640-996d-43e1-8e5d-e5ed2a514896" containerName="extract" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.541185 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-cp7lm" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.544056 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.544749 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-tfm8k" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.545001 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.570217 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-cp7lm"] Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.592160 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbrq7\" (UniqueName: \"kubernetes.io/projected/03bbe6ea-a40e-4d51-85fb-d85580cfe27f-kube-api-access-sbrq7\") pod \"cluster-logging-operator-ff9846bd-cp7lm\" (UID: \"03bbe6ea-a40e-4d51-85fb-d85580cfe27f\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-cp7lm" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.693415 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbrq7\" (UniqueName: \"kubernetes.io/projected/03bbe6ea-a40e-4d51-85fb-d85580cfe27f-kube-api-access-sbrq7\") pod \"cluster-logging-operator-ff9846bd-cp7lm\" (UID: \"03bbe6ea-a40e-4d51-85fb-d85580cfe27f\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-cp7lm" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.714352 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbrq7\" (UniqueName: \"kubernetes.io/projected/03bbe6ea-a40e-4d51-85fb-d85580cfe27f-kube-api-access-sbrq7\") pod \"cluster-logging-operator-ff9846bd-cp7lm\" (UID: \"03bbe6ea-a40e-4d51-85fb-d85580cfe27f\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-cp7lm" Nov 22 08:14:38 crc kubenswrapper[4735]: I1122 08:14:38.856745 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-cp7lm" Nov 22 08:14:39 crc kubenswrapper[4735]: I1122 08:14:39.353679 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-cp7lm"] Nov 22 08:14:40 crc kubenswrapper[4735]: I1122 08:14:40.279974 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-cp7lm" event={"ID":"03bbe6ea-a40e-4d51-85fb-d85580cfe27f","Type":"ContainerStarted","Data":"03406449f0689d2cd62bebbf9d7c4c894d53d762e0b60c2514c10d9bc1682d5e"} Nov 22 08:14:45 crc kubenswrapper[4735]: I1122 08:14:45.322311 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-cp7lm" event={"ID":"03bbe6ea-a40e-4d51-85fb-d85580cfe27f","Type":"ContainerStarted","Data":"1db77112850469cf50a24eb625df772e175b08d44f9496a8ee4afee3a7082419"} Nov 22 08:14:45 crc kubenswrapper[4735]: I1122 08:14:45.340798 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-cp7lm" podStartSLOduration=1.841040395 podStartE2EDuration="7.340772418s" podCreationTimestamp="2025-11-22 08:14:38 +0000 UTC" firstStartedPulling="2025-11-22 08:14:39.370511827 +0000 UTC m=+700.974850442" lastFinishedPulling="2025-11-22 08:14:44.87024387 +0000 UTC m=+706.474582465" observedRunningTime="2025-11-22 08:14:45.335881683 +0000 UTC m=+706.940220308" watchObservedRunningTime="2025-11-22 08:14:45.340772418 +0000 UTC m=+706.945111043" Nov 22 08:14:46 crc kubenswrapper[4735]: I1122 08:14:46.152214 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:14:46 crc kubenswrapper[4735]: I1122 08:14:46.152284 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.756961 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4"] Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.758580 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.760813 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.760880 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.761216 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-658bh" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.761212 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.761377 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.761778 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.787880 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4"] Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.872525 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-webhook-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.872590 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b0172333-c8a0-4e17-a312-d3362d0fe8da-manager-config\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.872652 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89f8w\" (UniqueName: \"kubernetes.io/projected/b0172333-c8a0-4e17-a312-d3362d0fe8da-kube-api-access-89f8w\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.872808 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.873452 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-apiservice-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.975307 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89f8w\" (UniqueName: \"kubernetes.io/projected/b0172333-c8a0-4e17-a312-d3362d0fe8da-kube-api-access-89f8w\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.975368 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.975451 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-apiservice-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.975491 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-webhook-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.975521 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b0172333-c8a0-4e17-a312-d3362d0fe8da-manager-config\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.976492 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b0172333-c8a0-4e17-a312-d3362d0fe8da-manager-config\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.987180 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.994048 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-apiservice-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:49 crc kubenswrapper[4735]: I1122 08:14:49.998439 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89f8w\" (UniqueName: \"kubernetes.io/projected/b0172333-c8a0-4e17-a312-d3362d0fe8da-kube-api-access-89f8w\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:50 crc kubenswrapper[4735]: I1122 08:14:50.002337 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0172333-c8a0-4e17-a312-d3362d0fe8da-webhook-cert\") pod \"loki-operator-controller-manager-5b4b6fd4bd-jc7r4\" (UID: \"b0172333-c8a0-4e17-a312-d3362d0fe8da\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:50 crc kubenswrapper[4735]: I1122 08:14:50.076613 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:14:50 crc kubenswrapper[4735]: I1122 08:14:50.657119 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4"] Nov 22 08:14:51 crc kubenswrapper[4735]: I1122 08:14:51.369694 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" event={"ID":"b0172333-c8a0-4e17-a312-d3362d0fe8da","Type":"ContainerStarted","Data":"32d934f045bfe5f2a4bc6cbe2ee3a2095636dd52ee4aaa0bdbf7b807cdda4ba9"} Nov 22 08:14:54 crc kubenswrapper[4735]: I1122 08:14:54.386798 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" event={"ID":"b0172333-c8a0-4e17-a312-d3362d0fe8da","Type":"ContainerStarted","Data":"d51ccc662ba949f23fa73704eafb26b4331651446140c40c25e6b716f5375691"} Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.144572 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp"] Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.145954 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.149894 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.150989 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.176485 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp"] Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.188875 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdp6v\" (UniqueName: \"kubernetes.io/projected/6d7eea70-9a64-4989-87a3-b4d6059cb49a-kube-api-access-gdp6v\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.188980 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7eea70-9a64-4989-87a3-b4d6059cb49a-config-volume\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.189018 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7eea70-9a64-4989-87a3-b4d6059cb49a-secret-volume\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.289850 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7eea70-9a64-4989-87a3-b4d6059cb49a-secret-volume\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.289946 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdp6v\" (UniqueName: \"kubernetes.io/projected/6d7eea70-9a64-4989-87a3-b4d6059cb49a-kube-api-access-gdp6v\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.289987 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7eea70-9a64-4989-87a3-b4d6059cb49a-config-volume\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.290736 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7eea70-9a64-4989-87a3-b4d6059cb49a-config-volume\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.295783 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7eea70-9a64-4989-87a3-b4d6059cb49a-secret-volume\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.310965 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdp6v\" (UniqueName: \"kubernetes.io/projected/6d7eea70-9a64-4989-87a3-b4d6059cb49a-kube-api-access-gdp6v\") pod \"collect-profiles-29396655-bxlgp\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:00 crc kubenswrapper[4735]: I1122 08:15:00.474851 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:01 crc kubenswrapper[4735]: I1122 08:15:01.445152 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" event={"ID":"b0172333-c8a0-4e17-a312-d3362d0fe8da","Type":"ContainerStarted","Data":"4e21e4b54b7fa9e2bd9358f6297eb063a38977987e8e77ff3c29250ea8dbd235"} Nov 22 08:15:01 crc kubenswrapper[4735]: I1122 08:15:01.445700 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:15:01 crc kubenswrapper[4735]: I1122 08:15:01.454365 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" Nov 22 08:15:01 crc kubenswrapper[4735]: I1122 08:15:01.464513 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp"] Nov 22 08:15:01 crc kubenswrapper[4735]: I1122 08:15:01.479384 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-5b4b6fd4bd-jc7r4" podStartSLOduration=2.044076659 podStartE2EDuration="12.479362201s" podCreationTimestamp="2025-11-22 08:14:49 +0000 UTC" firstStartedPulling="2025-11-22 08:14:50.664839678 +0000 UTC m=+712.269178283" lastFinishedPulling="2025-11-22 08:15:01.10012522 +0000 UTC m=+722.704463825" observedRunningTime="2025-11-22 08:15:01.472781572 +0000 UTC m=+723.077120177" watchObservedRunningTime="2025-11-22 08:15:01.479362201 +0000 UTC m=+723.083700806" Nov 22 08:15:02 crc kubenswrapper[4735]: I1122 08:15:02.452075 4735 generic.go:334] "Generic (PLEG): container finished" podID="6d7eea70-9a64-4989-87a3-b4d6059cb49a" containerID="02294add5e71d912d138c2807debf2933a4179cd8a1179a36e6c5dd982f94773" exitCode=0 Nov 22 08:15:02 crc kubenswrapper[4735]: I1122 08:15:02.452180 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" event={"ID":"6d7eea70-9a64-4989-87a3-b4d6059cb49a","Type":"ContainerDied","Data":"02294add5e71d912d138c2807debf2933a4179cd8a1179a36e6c5dd982f94773"} Nov 22 08:15:02 crc kubenswrapper[4735]: I1122 08:15:02.452388 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" event={"ID":"6d7eea70-9a64-4989-87a3-b4d6059cb49a","Type":"ContainerStarted","Data":"8bbf0bce0e161f751e9da6fe9953b6b2bbe872b2b74acf0fdfc208ae5037c2ad"} Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.730889 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.839827 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7eea70-9a64-4989-87a3-b4d6059cb49a-config-volume\") pod \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.839993 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdp6v\" (UniqueName: \"kubernetes.io/projected/6d7eea70-9a64-4989-87a3-b4d6059cb49a-kube-api-access-gdp6v\") pod \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.840042 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7eea70-9a64-4989-87a3-b4d6059cb49a-secret-volume\") pod \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\" (UID: \"6d7eea70-9a64-4989-87a3-b4d6059cb49a\") " Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.840839 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d7eea70-9a64-4989-87a3-b4d6059cb49a-config-volume" (OuterVolumeSpecName: "config-volume") pod "6d7eea70-9a64-4989-87a3-b4d6059cb49a" (UID: "6d7eea70-9a64-4989-87a3-b4d6059cb49a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.846730 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7eea70-9a64-4989-87a3-b4d6059cb49a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6d7eea70-9a64-4989-87a3-b4d6059cb49a" (UID: "6d7eea70-9a64-4989-87a3-b4d6059cb49a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.862661 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d7eea70-9a64-4989-87a3-b4d6059cb49a-kube-api-access-gdp6v" (OuterVolumeSpecName: "kube-api-access-gdp6v") pod "6d7eea70-9a64-4989-87a3-b4d6059cb49a" (UID: "6d7eea70-9a64-4989-87a3-b4d6059cb49a"). InnerVolumeSpecName "kube-api-access-gdp6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.941511 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7eea70-9a64-4989-87a3-b4d6059cb49a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.941550 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7eea70-9a64-4989-87a3-b4d6059cb49a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:03 crc kubenswrapper[4735]: I1122 08:15:03.941561 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdp6v\" (UniqueName: \"kubernetes.io/projected/6d7eea70-9a64-4989-87a3-b4d6059cb49a-kube-api-access-gdp6v\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:04 crc kubenswrapper[4735]: I1122 08:15:04.466323 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" event={"ID":"6d7eea70-9a64-4989-87a3-b4d6059cb49a","Type":"ContainerDied","Data":"8bbf0bce0e161f751e9da6fe9953b6b2bbe872b2b74acf0fdfc208ae5037c2ad"} Nov 22 08:15:04 crc kubenswrapper[4735]: I1122 08:15:04.466367 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bbf0bce0e161f751e9da6fe9953b6b2bbe872b2b74acf0fdfc208ae5037c2ad" Nov 22 08:15:04 crc kubenswrapper[4735]: I1122 08:15:04.466387 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.437381 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Nov 22 08:15:06 crc kubenswrapper[4735]: E1122 08:15:06.438808 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d7eea70-9a64-4989-87a3-b4d6059cb49a" containerName="collect-profiles" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.438842 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d7eea70-9a64-4989-87a3-b4d6059cb49a" containerName="collect-profiles" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.439199 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d7eea70-9a64-4989-87a3-b4d6059cb49a" containerName="collect-profiles" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.441872 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.446093 4735 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-2pq4j" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.446728 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.448176 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.461594 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.490137 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvrwj\" (UniqueName: \"kubernetes.io/projected/07fa7fd0-c626-446b-91f8-2df5e63629fe-kube-api-access-kvrwj\") pod \"minio\" (UID: \"07fa7fd0-c626-446b-91f8-2df5e63629fe\") " pod="minio-dev/minio" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.490215 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f49b929d-1506-460d-ab8b-957b210ba50a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49b929d-1506-460d-ab8b-957b210ba50a\") pod \"minio\" (UID: \"07fa7fd0-c626-446b-91f8-2df5e63629fe\") " pod="minio-dev/minio" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.591352 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f49b929d-1506-460d-ab8b-957b210ba50a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49b929d-1506-460d-ab8b-957b210ba50a\") pod \"minio\" (UID: \"07fa7fd0-c626-446b-91f8-2df5e63629fe\") " pod="minio-dev/minio" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.591517 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvrwj\" (UniqueName: \"kubernetes.io/projected/07fa7fd0-c626-446b-91f8-2df5e63629fe-kube-api-access-kvrwj\") pod \"minio\" (UID: \"07fa7fd0-c626-446b-91f8-2df5e63629fe\") " pod="minio-dev/minio" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.596327 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.596376 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f49b929d-1506-460d-ab8b-957b210ba50a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49b929d-1506-460d-ab8b-957b210ba50a\") pod \"minio\" (UID: \"07fa7fd0-c626-446b-91f8-2df5e63629fe\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b2bcbf681b248328aeec25b74cb9db59775fb233ffcb7990c6b853ecb1c017c4/globalmount\"" pod="minio-dev/minio" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.614255 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvrwj\" (UniqueName: \"kubernetes.io/projected/07fa7fd0-c626-446b-91f8-2df5e63629fe-kube-api-access-kvrwj\") pod \"minio\" (UID: \"07fa7fd0-c626-446b-91f8-2df5e63629fe\") " pod="minio-dev/minio" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.672615 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f49b929d-1506-460d-ab8b-957b210ba50a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49b929d-1506-460d-ab8b-957b210ba50a\") pod \"minio\" (UID: \"07fa7fd0-c626-446b-91f8-2df5e63629fe\") " pod="minio-dev/minio" Nov 22 08:15:06 crc kubenswrapper[4735]: I1122 08:15:06.779656 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 22 08:15:07 crc kubenswrapper[4735]: I1122 08:15:07.219761 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 22 08:15:07 crc kubenswrapper[4735]: I1122 08:15:07.498189 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"07fa7fd0-c626-446b-91f8-2df5e63629fe","Type":"ContainerStarted","Data":"3b07073675d4412e3664134492df36d40f9d47e922d6b084c3d85aecc324f4f8"} Nov 22 08:15:14 crc kubenswrapper[4735]: I1122 08:15:14.541888 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"07fa7fd0-c626-446b-91f8-2df5e63629fe","Type":"ContainerStarted","Data":"d6da16d8181f95f37fe6ad818ffbe8bdabeeffa40a4fc7efeec02353d3b8de14"} Nov 22 08:15:14 crc kubenswrapper[4735]: I1122 08:15:14.556989 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.260994441 podStartE2EDuration="10.556972409s" podCreationTimestamp="2025-11-22 08:15:04 +0000 UTC" firstStartedPulling="2025-11-22 08:15:07.230325654 +0000 UTC m=+728.834664259" lastFinishedPulling="2025-11-22 08:15:13.526303622 +0000 UTC m=+735.130642227" observedRunningTime="2025-11-22 08:15:14.554148181 +0000 UTC m=+736.158486786" watchObservedRunningTime="2025-11-22 08:15:14.556972409 +0000 UTC m=+736.161311014" Nov 22 08:15:16 crc kubenswrapper[4735]: I1122 08:15:16.131980 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:15:16 crc kubenswrapper[4735]: I1122 08:15:16.132272 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.631067 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx"] Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.631892 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.634713 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.634932 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.638753 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.640668 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.640725 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-w4cmv" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.655054 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx"] Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.782554 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.782683 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.782746 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bab3bec-f764-4c68-a487-6d7dbc40e258-config\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.782788 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwtbx\" (UniqueName: \"kubernetes.io/projected/2bab3bec-f764-4c68-a487-6d7dbc40e258-kube-api-access-fwtbx\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.782836 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.799082 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-zhskj"] Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.799841 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.804844 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.804860 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.804990 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.818243 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-zhskj"] Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.884612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.884666 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bab3bec-f764-4c68-a487-6d7dbc40e258-config\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.884694 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwtbx\" (UniqueName: \"kubernetes.io/projected/2bab3bec-f764-4c68-a487-6d7dbc40e258-kube-api-access-fwtbx\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.884722 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.884771 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.885731 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bab3bec-f764-4c68-a487-6d7dbc40e258-config\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.885826 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.899786 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg"] Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.900616 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.903408 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.903766 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.903847 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.906770 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/2bab3bec-f764-4c68-a487-6d7dbc40e258-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.919220 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg"] Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.924207 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwtbx\" (UniqueName: \"kubernetes.io/projected/2bab3bec-f764-4c68-a487-6d7dbc40e258-kube-api-access-fwtbx\") pod \"logging-loki-distributor-76cc67bf56-hvxhx\" (UID: \"2bab3bec-f764-4c68-a487-6d7dbc40e258\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.948354 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.986449 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.986522 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.986556 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.986622 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35c00115-3277-4e70-ae42-9a13e4a4459e-config\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.986703 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtld5\" (UniqueName: \"kubernetes.io/projected/35c00115-3277-4e70-ae42-9a13e4a4459e-kube-api-access-qtld5\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:18 crc kubenswrapper[4735]: I1122 08:15:18.986726 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.044706 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.053215 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.058776 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.058821 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.058779 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.058921 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.059846 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.064533 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.075685 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.078057 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.090624 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-9f5gm" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092195 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-lokistack-gateway\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092234 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtld5\" (UniqueName: \"kubernetes.io/projected/35c00115-3277-4e70-ae42-9a13e4a4459e-kube-api-access-qtld5\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092272 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092304 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092350 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092366 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrccf\" (UniqueName: \"kubernetes.io/projected/5caa9595-315a-40a7-a46f-5e6d4b13c320-kube-api-access-lrccf\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092391 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092410 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092432 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlph7\" (UniqueName: \"kubernetes.io/projected/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-kube-api-access-tlph7\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092451 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tenants\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092487 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-rbac\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092502 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tls-secret\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092521 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-rbac\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092536 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092553 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tls-secret\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092580 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092598 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092616 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092634 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092652 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-config\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092667 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092689 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092707 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-lokistack-gateway\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092724 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092744 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35c00115-3277-4e70-ae42-9a13e4a4459e-config\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092760 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tenants\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.092774 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9jkz\" (UniqueName: \"kubernetes.io/projected/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-kube-api-access-c9jkz\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.097199 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.099306 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35c00115-3277-4e70-ae42-9a13e4a4459e-config\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.100421 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.101965 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.105710 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/35c00115-3277-4e70-ae42-9a13e4a4459e-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.106967 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.113975 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtld5\" (UniqueName: \"kubernetes.io/projected/35c00115-3277-4e70-ae42-9a13e4a4459e-kube-api-access-qtld5\") pod \"logging-loki-querier-5895d59bb8-zhskj\" (UID: \"35c00115-3277-4e70-ae42-9a13e4a4459e\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.121918 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194004 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-lokistack-gateway\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194050 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194090 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194108 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrccf\" (UniqueName: \"kubernetes.io/projected/5caa9595-315a-40a7-a46f-5e6d4b13c320-kube-api-access-lrccf\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194135 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194152 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194172 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlph7\" (UniqueName: \"kubernetes.io/projected/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-kube-api-access-tlph7\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194190 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tenants\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194206 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tls-secret\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194225 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-rbac\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194246 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-rbac\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194264 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194280 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tls-secret\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194306 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194325 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-config\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194343 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194363 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194382 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-lokistack-gateway\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194399 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194415 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9jkz\" (UniqueName: \"kubernetes.io/projected/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-kube-api-access-c9jkz\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.194430 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tenants\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.196769 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-rbac\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.197437 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-lokistack-gateway\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.198170 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.198373 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tenants\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.198812 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.199110 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-rbac\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.205292 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: E1122 08:15:19.205392 4735 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Nov 22 08:15:19 crc kubenswrapper[4735]: E1122 08:15:19.205433 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tls-secret podName:3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0 nodeName:}" failed. No retries permitted until 2025-11-22 08:15:19.705419778 +0000 UTC m=+741.309758383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tls-secret") pod "logging-loki-gateway-6f9c8b877-jv5cr" (UID: "3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0") : secret "logging-loki-gateway-http" not found Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.206220 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.207444 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-lokistack-gateway\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.208109 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.210769 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-config\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.211296 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.211697 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: E1122 08:15:19.211831 4735 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Nov 22 08:15:19 crc kubenswrapper[4735]: E1122 08:15:19.211898 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tls-secret podName:5caa9595-315a-40a7-a46f-5e6d4b13c320 nodeName:}" failed. No retries permitted until 2025-11-22 08:15:19.711879472 +0000 UTC m=+741.316218077 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tls-secret") pod "logging-loki-gateway-6f9c8b877-5dtcw" (UID: "5caa9595-315a-40a7-a46f-5e6d4b13c320") : secret "logging-loki-gateway-http" not found Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.215274 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.216494 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.228909 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrccf\" (UniqueName: \"kubernetes.io/projected/5caa9595-315a-40a7-a46f-5e6d4b13c320-kube-api-access-lrccf\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.232381 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tenants\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.232722 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9jkz\" (UniqueName: \"kubernetes.io/projected/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-kube-api-access-c9jkz\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.236415 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlph7\" (UniqueName: \"kubernetes.io/projected/1b08c63f-3828-4fd2-95a8-7ef002d5fdbd-kube-api-access-tlph7\") pod \"logging-loki-query-frontend-84558f7c9f-hxcvg\" (UID: \"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.270168 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.509684 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx"] Nov 22 08:15:19 crc kubenswrapper[4735]: W1122 08:15:19.514146 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bab3bec_f764_4c68_a487_6d7dbc40e258.slice/crio-5510f2fe64cb4f17e55e5749cc69cf6f67f4a1acf780ca5bafb902900545d65e WatchSource:0}: Error finding container 5510f2fe64cb4f17e55e5749cc69cf6f67f4a1acf780ca5bafb902900545d65e: Status 404 returned error can't find the container with id 5510f2fe64cb4f17e55e5749cc69cf6f67f4a1acf780ca5bafb902900545d65e Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.572759 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" event={"ID":"2bab3bec-f764-4c68-a487-6d7dbc40e258","Type":"ContainerStarted","Data":"5510f2fe64cb4f17e55e5749cc69cf6f67f4a1acf780ca5bafb902900545d65e"} Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.606315 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-zhskj"] Nov 22 08:15:19 crc kubenswrapper[4735]: W1122 08:15:19.609837 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35c00115_3277_4e70_ae42_9a13e4a4459e.slice/crio-776d12102eb1d4c2c5a49da773c2822a93c34f59f69973b7757ae87413343cae WatchSource:0}: Error finding container 776d12102eb1d4c2c5a49da773c2822a93c34f59f69973b7757ae87413343cae: Status 404 returned error can't find the container with id 776d12102eb1d4c2c5a49da773c2822a93c34f59f69973b7757ae87413343cae Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.749357 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg"] Nov 22 08:15:19 crc kubenswrapper[4735]: W1122 08:15:19.751543 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b08c63f_3828_4fd2_95a8_7ef002d5fdbd.slice/crio-ec6961a1fc23ca72cf315619d66660e3935952d148d391e465bf1e72c4e1a166 WatchSource:0}: Error finding container ec6961a1fc23ca72cf315619d66660e3935952d148d391e465bf1e72c4e1a166: Status 404 returned error can't find the container with id ec6961a1fc23ca72cf315619d66660e3935952d148d391e465bf1e72c4e1a166 Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.796362 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.798223 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.801316 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.801687 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.804701 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.806260 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tls-secret\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.806310 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tls-secret\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.811743 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/5caa9595-315a-40a7-a46f-5e6d4b13c320-tls-secret\") pod \"logging-loki-gateway-6f9c8b877-5dtcw\" (UID: \"5caa9595-315a-40a7-a46f-5e6d4b13c320\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.813094 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0-tls-secret\") pod \"logging-loki-gateway-6f9c8b877-jv5cr\" (UID: \"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0\") " pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.879788 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.880761 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.883237 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.883398 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.887377 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.911695 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.911765 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.911814 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3b1eee2-6b00-45b3-95bd-9e3005004633-config\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.911851 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8z8l\" (UniqueName: \"kubernetes.io/projected/d3b1eee2-6b00-45b3-95bd-9e3005004633-kube-api-access-h8z8l\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.911901 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-52cfb90f-1588-4167-9558-f8b56f1c4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-52cfb90f-1588-4167-9558-f8b56f1c4606\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.911940 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.911988 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7352ea09-b04a-46fd-a2df-b3bc5dde887a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7352ea09-b04a-46fd-a2df-b3bc5dde887a\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.912026 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.975824 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.979418 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.990607 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.996646 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Nov 22 08:15:19 crc kubenswrapper[4735]: I1122 08:15:19.996869 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.004954 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.009701 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014622 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8z8l\" (UniqueName: \"kubernetes.io/projected/d3b1eee2-6b00-45b3-95bd-9e3005004633-kube-api-access-h8z8l\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014688 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-52cfb90f-1588-4167-9558-f8b56f1c4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-52cfb90f-1588-4167-9558-f8b56f1c4606\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014723 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014752 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014781 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014816 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0a12d857-d9cd-466b-9164-704d77585948\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0a12d857-d9cd-466b-9164-704d77585948\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014857 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckj4g\" (UniqueName: \"kubernetes.io/projected/77078529-6c53-48d6-9bfc-671f3e2eba43-kube-api-access-ckj4g\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014885 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7352ea09-b04a-46fd-a2df-b3bc5dde887a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7352ea09-b04a-46fd-a2df-b3bc5dde887a\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014910 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014937 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.014965 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.015012 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.015057 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77078529-6c53-48d6-9bfc-671f3e2eba43-config\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.015086 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.015116 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3b1eee2-6b00-45b3-95bd-9e3005004633-config\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.016331 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3b1eee2-6b00-45b3-95bd-9e3005004633-config\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.018065 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.018111 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-52cfb90f-1588-4167-9558-f8b56f1c4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-52cfb90f-1588-4167-9558-f8b56f1c4606\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1df6935832ddf60dc3b9904bdebfaa68e363259a0de4811344f25812ec0e6343/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.022329 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.024044 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.031087 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.050539 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d3b1eee2-6b00-45b3-95bd-9e3005004633-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.058981 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.059033 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7352ea09-b04a-46fd-a2df-b3bc5dde887a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7352ea09-b04a-46fd-a2df-b3bc5dde887a\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ed09cefb9c28689855d1e33e4e60994381c9f2494a5a77d67a925ac5467234fa/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.100371 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8z8l\" (UniqueName: \"kubernetes.io/projected/d3b1eee2-6b00-45b3-95bd-9e3005004633-kube-api-access-h8z8l\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116395 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77078529-6c53-48d6-9bfc-671f3e2eba43-config\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116486 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116561 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckj4g\" (UniqueName: \"kubernetes.io/projected/77078529-6c53-48d6-9bfc-671f3e2eba43-kube-api-access-ckj4g\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116584 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116608 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116634 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116746 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zfjv\" (UniqueName: \"kubernetes.io/projected/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-kube-api-access-8zfjv\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116840 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116889 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116914 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116939 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.116977 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0a12d857-d9cd-466b-9164-704d77585948\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0a12d857-d9cd-466b-9164-704d77585948\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.117044 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-config\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.117072 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-490c2d52-64a5-4148-a0dc-6f9e19539b3a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-490c2d52-64a5-4148-a0dc-6f9e19539b3a\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.119228 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77078529-6c53-48d6-9bfc-671f3e2eba43-config\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.123741 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.125075 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.143296 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.144768 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/77078529-6c53-48d6-9bfc-671f3e2eba43-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.147080 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.147109 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0a12d857-d9cd-466b-9164-704d77585948\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0a12d857-d9cd-466b-9164-704d77585948\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2301d079ffa276d8fb54845b2f60611d787ee3f3973735cde86a7c6da5bc2ab6/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.156116 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckj4g\" (UniqueName: \"kubernetes.io/projected/77078529-6c53-48d6-9bfc-671f3e2eba43-kube-api-access-ckj4g\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.158300 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-52cfb90f-1588-4167-9558-f8b56f1c4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-52cfb90f-1588-4167-9558-f8b56f1c4606\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.166705 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7352ea09-b04a-46fd-a2df-b3bc5dde887a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7352ea09-b04a-46fd-a2df-b3bc5dde887a\") pod \"logging-loki-ingester-0\" (UID: \"d3b1eee2-6b00-45b3-95bd-9e3005004633\") " pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.179304 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0a12d857-d9cd-466b-9164-704d77585948\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0a12d857-d9cd-466b-9164-704d77585948\") pod \"logging-loki-compactor-0\" (UID: \"77078529-6c53-48d6-9bfc-671f3e2eba43\") " pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.218365 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.218416 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zfjv\" (UniqueName: \"kubernetes.io/projected/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-kube-api-access-8zfjv\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.218446 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.218495 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.218530 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-config\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.218547 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-490c2d52-64a5-4148-a0dc-6f9e19539b3a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-490c2d52-64a5-4148-a0dc-6f9e19539b3a\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.218589 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.221853 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.222302 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.223251 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-config\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.225898 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.227736 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.228877 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.228914 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-490c2d52-64a5-4148-a0dc-6f9e19539b3a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-490c2d52-64a5-4148-a0dc-6f9e19539b3a\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dc7d5537e8268d14ee204c8e273b59fe92797c31adcc3c6a5cbc82ab9cc98ca8/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.238346 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zfjv\" (UniqueName: \"kubernetes.io/projected/cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f-kube-api-access-8zfjv\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.251030 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.263578 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-490c2d52-64a5-4148-a0dc-6f9e19539b3a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-490c2d52-64a5-4148-a0dc-6f9e19539b3a\") pod \"logging-loki-index-gateway-0\" (UID: \"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.386560 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr"] Nov 22 08:15:20 crc kubenswrapper[4735]: W1122 08:15:20.396406 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fc1eb4e_6044_4fe5_a6ad_ad3dc18892a0.slice/crio-ba05746432b9d9854a4130b038a9dbed8ba610efc2b6d44d1b83dd1a67fe31ba WatchSource:0}: Error finding container ba05746432b9d9854a4130b038a9dbed8ba610efc2b6d44d1b83dd1a67fe31ba: Status 404 returned error can't find the container with id ba05746432b9d9854a4130b038a9dbed8ba610efc2b6d44d1b83dd1a67fe31ba Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.423228 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.444795 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.535433 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw"] Nov 22 08:15:20 crc kubenswrapper[4735]: W1122 08:15:20.544208 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5caa9595_315a_40a7_a46f_5e6d4b13c320.slice/crio-f9302258e3067a604419ca8e000f1caf82c2a485fa65ef672f17503cb10bf227 WatchSource:0}: Error finding container f9302258e3067a604419ca8e000f1caf82c2a485fa65ef672f17503cb10bf227: Status 404 returned error can't find the container with id f9302258e3067a604419ca8e000f1caf82c2a485fa65ef672f17503cb10bf227 Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.580041 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" event={"ID":"35c00115-3277-4e70-ae42-9a13e4a4459e","Type":"ContainerStarted","Data":"776d12102eb1d4c2c5a49da773c2822a93c34f59f69973b7757ae87413343cae"} Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.582981 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" event={"ID":"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0","Type":"ContainerStarted","Data":"ba05746432b9d9854a4130b038a9dbed8ba610efc2b6d44d1b83dd1a67fe31ba"} Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.584097 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" event={"ID":"5caa9595-315a-40a7-a46f-5e6d4b13c320","Type":"ContainerStarted","Data":"f9302258e3067a604419ca8e000f1caf82c2a485fa65ef672f17503cb10bf227"} Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.585265 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" event={"ID":"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd","Type":"ContainerStarted","Data":"ec6961a1fc23ca72cf315619d66660e3935952d148d391e465bf1e72c4e1a166"} Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.660950 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 22 08:15:20 crc kubenswrapper[4735]: W1122 08:15:20.669750 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77078529_6c53_48d6_9bfc_671f3e2eba43.slice/crio-7c8e39adfdf0124734263632c8592243fadcc20b48113c1f6adb6bc53084ef2c WatchSource:0}: Error finding container 7c8e39adfdf0124734263632c8592243fadcc20b48113c1f6adb6bc53084ef2c: Status 404 returned error can't find the container with id 7c8e39adfdf0124734263632c8592243fadcc20b48113c1f6adb6bc53084ef2c Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.823386 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 22 08:15:20 crc kubenswrapper[4735]: W1122 08:15:20.826979 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3b1eee2_6b00_45b3_95bd_9e3005004633.slice/crio-d474377eb5ef5b69b9977e67320d19eee4257b7f9453b05b67075c657c830108 WatchSource:0}: Error finding container d474377eb5ef5b69b9977e67320d19eee4257b7f9453b05b67075c657c830108: Status 404 returned error can't find the container with id d474377eb5ef5b69b9977e67320d19eee4257b7f9453b05b67075c657c830108 Nov 22 08:15:20 crc kubenswrapper[4735]: I1122 08:15:20.882662 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 22 08:15:20 crc kubenswrapper[4735]: W1122 08:15:20.892216 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd0c04d0_3cae_4611_9cdb_57e95cbb3a5f.slice/crio-ca394394a9a20dcd2cf23d436d2e388a1465f13e921ad6af2ccc53689783f8e4 WatchSource:0}: Error finding container ca394394a9a20dcd2cf23d436d2e388a1465f13e921ad6af2ccc53689783f8e4: Status 404 returned error can't find the container with id ca394394a9a20dcd2cf23d436d2e388a1465f13e921ad6af2ccc53689783f8e4 Nov 22 08:15:21 crc kubenswrapper[4735]: I1122 08:15:21.594025 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"77078529-6c53-48d6-9bfc-671f3e2eba43","Type":"ContainerStarted","Data":"7c8e39adfdf0124734263632c8592243fadcc20b48113c1f6adb6bc53084ef2c"} Nov 22 08:15:21 crc kubenswrapper[4735]: I1122 08:15:21.595549 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"d3b1eee2-6b00-45b3-95bd-9e3005004633","Type":"ContainerStarted","Data":"d474377eb5ef5b69b9977e67320d19eee4257b7f9453b05b67075c657c830108"} Nov 22 08:15:21 crc kubenswrapper[4735]: I1122 08:15:21.596839 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f","Type":"ContainerStarted","Data":"ca394394a9a20dcd2cf23d436d2e388a1465f13e921ad6af2ccc53689783f8e4"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.616693 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"d3b1eee2-6b00-45b3-95bd-9e3005004633","Type":"ContainerStarted","Data":"c57e304636383632186b18170f1380c7fc3feca17d4fe45671b61a7a791d8bc3"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.616995 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.618135 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f","Type":"ContainerStarted","Data":"74dd8e964295fbeeaa307c861999d538e20702d199abb0a3164c0d8945f8b17b"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.618203 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.619779 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" event={"ID":"35c00115-3277-4e70-ae42-9a13e4a4459e","Type":"ContainerStarted","Data":"15bdb86753488307597706411b2c3b4316416787912c1fec40f7e939bb9848d5"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.619910 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.648185 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"77078529-6c53-48d6-9bfc-671f3e2eba43","Type":"ContainerStarted","Data":"a5f86e7950c8b7ee50ccfa76e72b6f1c2c903da9da7b202bc9469a5be03508af"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.648294 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.650475 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" event={"ID":"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0","Type":"ContainerStarted","Data":"050da6364e030c0449f1738cd8b39bff2b5faa00b5ccdc5bef9a184672ca09fd"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.651570 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" event={"ID":"5caa9595-315a-40a7-a46f-5e6d4b13c320","Type":"ContainerStarted","Data":"8b96a6d8af500d8518e3161c8b9d3fdbb94ca875f124453d71276a647bd49406"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.653088 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.8264333539999997 podStartE2EDuration="6.653064531s" podCreationTimestamp="2025-11-22 08:15:18 +0000 UTC" firstStartedPulling="2025-11-22 08:15:20.831043866 +0000 UTC m=+742.435382471" lastFinishedPulling="2025-11-22 08:15:23.657675023 +0000 UTC m=+745.262013648" observedRunningTime="2025-11-22 08:15:24.642141276 +0000 UTC m=+746.246479951" watchObservedRunningTime="2025-11-22 08:15:24.653064531 +0000 UTC m=+746.257403176" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.656848 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" event={"ID":"1b08c63f-3828-4fd2-95a8-7ef002d5fdbd","Type":"ContainerStarted","Data":"e320f3215b722c6f9c2402af8c33d58b34a4700e39d62d801a2b03e606e8c548"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.657020 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.660328 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" event={"ID":"2bab3bec-f764-4c68-a487-6d7dbc40e258","Type":"ContainerStarted","Data":"7ea4893001e6d03205becb53129a1070bd48080149b8a61bb8700a6aa349e83d"} Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.660517 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.673589 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.922908824 podStartE2EDuration="6.673565293s" podCreationTimestamp="2025-11-22 08:15:18 +0000 UTC" firstStartedPulling="2025-11-22 08:15:20.895210805 +0000 UTC m=+742.499549410" lastFinishedPulling="2025-11-22 08:15:23.645867234 +0000 UTC m=+745.250205879" observedRunningTime="2025-11-22 08:15:24.670841977 +0000 UTC m=+746.275180582" watchObservedRunningTime="2025-11-22 08:15:24.673565293 +0000 UTC m=+746.277903898" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.694995 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" podStartSLOduration=2.641807719 podStartE2EDuration="6.694973589s" podCreationTimestamp="2025-11-22 08:15:18 +0000 UTC" firstStartedPulling="2025-11-22 08:15:19.612580298 +0000 UTC m=+741.216918903" lastFinishedPulling="2025-11-22 08:15:23.665746138 +0000 UTC m=+745.270084773" observedRunningTime="2025-11-22 08:15:24.690432603 +0000 UTC m=+746.294771218" watchObservedRunningTime="2025-11-22 08:15:24.694973589 +0000 UTC m=+746.299312194" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.724725 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" podStartSLOduration=2.83237432 podStartE2EDuration="6.724710878s" podCreationTimestamp="2025-11-22 08:15:18 +0000 UTC" firstStartedPulling="2025-11-22 08:15:19.753561567 +0000 UTC m=+741.357900172" lastFinishedPulling="2025-11-22 08:15:23.645898125 +0000 UTC m=+745.250236730" observedRunningTime="2025-11-22 08:15:24.720930463 +0000 UTC m=+746.325269068" watchObservedRunningTime="2025-11-22 08:15:24.724710878 +0000 UTC m=+746.329049473" Nov 22 08:15:24 crc kubenswrapper[4735]: I1122 08:15:24.742747 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=4.089849788 podStartE2EDuration="6.74273127s" podCreationTimestamp="2025-11-22 08:15:18 +0000 UTC" firstStartedPulling="2025-11-22 08:15:20.672701803 +0000 UTC m=+742.277040408" lastFinishedPulling="2025-11-22 08:15:23.325583285 +0000 UTC m=+744.929921890" observedRunningTime="2025-11-22 08:15:24.738965826 +0000 UTC m=+746.343304431" watchObservedRunningTime="2025-11-22 08:15:24.74273127 +0000 UTC m=+746.347069875" Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.230029 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" podStartSLOduration=5.157059094 podStartE2EDuration="9.230004927s" podCreationTimestamp="2025-11-22 08:15:18 +0000 UTC" firstStartedPulling="2025-11-22 08:15:19.518648587 +0000 UTC m=+741.122987192" lastFinishedPulling="2025-11-22 08:15:23.59159438 +0000 UTC m=+745.195933025" observedRunningTime="2025-11-22 08:15:24.772940083 +0000 UTC m=+746.377278688" watchObservedRunningTime="2025-11-22 08:15:27.230004927 +0000 UTC m=+748.834343552" Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.236915 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zqccv"] Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.237225 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" podUID="e095b65e-6f13-4a89-8418-78f5bda82cb5" containerName="controller-manager" containerID="cri-o://fccf5d6f15567cd849c87eedb7e31d04969c435541bbc3011e279c73ed013406" gracePeriod=30 Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.249771 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4"] Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.250038 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" podUID="2da8d731-f76e-472f-bb05-b2242fb9a652" containerName="route-controller-manager" containerID="cri-o://e377a3308baafe60f81aa8f63d7e6d2c6553997e36c10002f44b98cb04b05cb0" gracePeriod=30 Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.689516 4735 generic.go:334] "Generic (PLEG): container finished" podID="2da8d731-f76e-472f-bb05-b2242fb9a652" containerID="e377a3308baafe60f81aa8f63d7e6d2c6553997e36c10002f44b98cb04b05cb0" exitCode=0 Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.689618 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" event={"ID":"2da8d731-f76e-472f-bb05-b2242fb9a652","Type":"ContainerDied","Data":"e377a3308baafe60f81aa8f63d7e6d2c6553997e36c10002f44b98cb04b05cb0"} Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.691375 4735 generic.go:334] "Generic (PLEG): container finished" podID="e095b65e-6f13-4a89-8418-78f5bda82cb5" containerID="fccf5d6f15567cd849c87eedb7e31d04969c435541bbc3011e279c73ed013406" exitCode=0 Nov 22 08:15:27 crc kubenswrapper[4735]: I1122 08:15:27.691472 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" event={"ID":"e095b65e-6f13-4a89-8418-78f5bda82cb5","Type":"ContainerDied","Data":"fccf5d6f15567cd849c87eedb7e31d04969c435541bbc3011e279c73ed013406"} Nov 22 08:15:28 crc kubenswrapper[4735]: I1122 08:15:28.938867 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:15:28 crc kubenswrapper[4735]: I1122 08:15:28.944092 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:15:28 crc kubenswrapper[4735]: I1122 08:15:28.972884 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl"] Nov 22 08:15:28 crc kubenswrapper[4735]: E1122 08:15:28.973221 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da8d731-f76e-472f-bb05-b2242fb9a652" containerName="route-controller-manager" Nov 22 08:15:28 crc kubenswrapper[4735]: I1122 08:15:28.973243 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da8d731-f76e-472f-bb05-b2242fb9a652" containerName="route-controller-manager" Nov 22 08:15:28 crc kubenswrapper[4735]: E1122 08:15:28.973280 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e095b65e-6f13-4a89-8418-78f5bda82cb5" containerName="controller-manager" Nov 22 08:15:28 crc kubenswrapper[4735]: I1122 08:15:28.973290 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e095b65e-6f13-4a89-8418-78f5bda82cb5" containerName="controller-manager" Nov 22 08:15:28 crc kubenswrapper[4735]: I1122 08:15:28.973440 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da8d731-f76e-472f-bb05-b2242fb9a652" containerName="route-controller-manager" Nov 22 08:15:28 crc kubenswrapper[4735]: I1122 08:15:28.973473 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e095b65e-6f13-4a89-8418-78f5bda82cb5" containerName="controller-manager" Nov 22 08:15:28 crc kubenswrapper[4735]: I1122 08:15:28.974061 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996488 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2da8d731-f76e-472f-bb05-b2242fb9a652-serving-cert\") pod \"2da8d731-f76e-472f-bb05-b2242fb9a652\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996549 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6hhs\" (UniqueName: \"kubernetes.io/projected/2da8d731-f76e-472f-bb05-b2242fb9a652-kube-api-access-n6hhs\") pod \"2da8d731-f76e-472f-bb05-b2242fb9a652\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996591 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-config\") pod \"2da8d731-f76e-472f-bb05-b2242fb9a652\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996626 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-client-ca\") pod \"2da8d731-f76e-472f-bb05-b2242fb9a652\" (UID: \"2da8d731-f76e-472f-bb05-b2242fb9a652\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996693 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlfq7\" (UniqueName: \"kubernetes.io/projected/e095b65e-6f13-4a89-8418-78f5bda82cb5-kube-api-access-dlfq7\") pod \"e095b65e-6f13-4a89-8418-78f5bda82cb5\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996717 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-config\") pod \"e095b65e-6f13-4a89-8418-78f5bda82cb5\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996738 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e095b65e-6f13-4a89-8418-78f5bda82cb5-serving-cert\") pod \"e095b65e-6f13-4a89-8418-78f5bda82cb5\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996780 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-proxy-ca-bundles\") pod \"e095b65e-6f13-4a89-8418-78f5bda82cb5\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.996818 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-client-ca\") pod \"e095b65e-6f13-4a89-8418-78f5bda82cb5\" (UID: \"e095b65e-6f13-4a89-8418-78f5bda82cb5\") " Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.997109 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ee43fb-95a1-40e2-837c-38f8e73c0357-config\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.997161 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cfmv\" (UniqueName: \"kubernetes.io/projected/a2ee43fb-95a1-40e2-837c-38f8e73c0357-kube-api-access-4cfmv\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.997209 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2ee43fb-95a1-40e2-837c-38f8e73c0357-client-ca\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:28.997248 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ee43fb-95a1-40e2-837c-38f8e73c0357-serving-cert\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.001276 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-client-ca" (OuterVolumeSpecName: "client-ca") pod "2da8d731-f76e-472f-bb05-b2242fb9a652" (UID: "2da8d731-f76e-472f-bb05-b2242fb9a652"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.002149 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-config" (OuterVolumeSpecName: "config") pod "2da8d731-f76e-472f-bb05-b2242fb9a652" (UID: "2da8d731-f76e-472f-bb05-b2242fb9a652"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.002767 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-config" (OuterVolumeSpecName: "config") pod "e095b65e-6f13-4a89-8418-78f5bda82cb5" (UID: "e095b65e-6f13-4a89-8418-78f5bda82cb5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.003773 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-client-ca" (OuterVolumeSpecName: "client-ca") pod "e095b65e-6f13-4a89-8418-78f5bda82cb5" (UID: "e095b65e-6f13-4a89-8418-78f5bda82cb5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.003944 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e095b65e-6f13-4a89-8418-78f5bda82cb5" (UID: "e095b65e-6f13-4a89-8418-78f5bda82cb5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.010608 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl"] Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.027348 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e095b65e-6f13-4a89-8418-78f5bda82cb5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e095b65e-6f13-4a89-8418-78f5bda82cb5" (UID: "e095b65e-6f13-4a89-8418-78f5bda82cb5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.027777 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da8d731-f76e-472f-bb05-b2242fb9a652-kube-api-access-n6hhs" (OuterVolumeSpecName: "kube-api-access-n6hhs") pod "2da8d731-f76e-472f-bb05-b2242fb9a652" (UID: "2da8d731-f76e-472f-bb05-b2242fb9a652"). InnerVolumeSpecName "kube-api-access-n6hhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.028195 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e095b65e-6f13-4a89-8418-78f5bda82cb5-kube-api-access-dlfq7" (OuterVolumeSpecName: "kube-api-access-dlfq7") pod "e095b65e-6f13-4a89-8418-78f5bda82cb5" (UID: "e095b65e-6f13-4a89-8418-78f5bda82cb5"). InnerVolumeSpecName "kube-api-access-dlfq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.028521 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da8d731-f76e-472f-bb05-b2242fb9a652-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2da8d731-f76e-472f-bb05-b2242fb9a652" (UID: "2da8d731-f76e-472f-bb05-b2242fb9a652"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.098751 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ee43fb-95a1-40e2-837c-38f8e73c0357-config\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.098826 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cfmv\" (UniqueName: \"kubernetes.io/projected/a2ee43fb-95a1-40e2-837c-38f8e73c0357-kube-api-access-4cfmv\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.098875 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2ee43fb-95a1-40e2-837c-38f8e73c0357-client-ca\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.098911 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ee43fb-95a1-40e2-837c-38f8e73c0357-serving-cert\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099000 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2da8d731-f76e-472f-bb05-b2242fb9a652-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099017 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6hhs\" (UniqueName: \"kubernetes.io/projected/2da8d731-f76e-472f-bb05-b2242fb9a652-kube-api-access-n6hhs\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099029 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099042 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2da8d731-f76e-472f-bb05-b2242fb9a652-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099054 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlfq7\" (UniqueName: \"kubernetes.io/projected/e095b65e-6f13-4a89-8418-78f5bda82cb5-kube-api-access-dlfq7\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099069 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099080 4735 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e095b65e-6f13-4a89-8418-78f5bda82cb5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099091 4735 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.099102 4735 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e095b65e-6f13-4a89-8418-78f5bda82cb5-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.100215 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ee43fb-95a1-40e2-837c-38f8e73c0357-config\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.100324 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2ee43fb-95a1-40e2-837c-38f8e73c0357-client-ca\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.103166 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ee43fb-95a1-40e2-837c-38f8e73c0357-serving-cert\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.123341 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cfmv\" (UniqueName: \"kubernetes.io/projected/a2ee43fb-95a1-40e2-837c-38f8e73c0357-kube-api-access-4cfmv\") pod \"route-controller-manager-84bd655b67-d5jgl\" (UID: \"a2ee43fb-95a1-40e2-837c-38f8e73c0357\") " pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.291814 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.618322 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl"] Nov 22 08:15:29 crc kubenswrapper[4735]: W1122 08:15:29.623916 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2ee43fb_95a1_40e2_837c_38f8e73c0357.slice/crio-42cac2f350a987fce84f2602f43e63e77740d21ba48f9c58948df0fa189254fc WatchSource:0}: Error finding container 42cac2f350a987fce84f2602f43e63e77740d21ba48f9c58948df0fa189254fc: Status 404 returned error can't find the container with id 42cac2f350a987fce84f2602f43e63e77740d21ba48f9c58948df0fa189254fc Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.706373 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" event={"ID":"5caa9595-315a-40a7-a46f-5e6d4b13c320","Type":"ContainerStarted","Data":"34ec6dcf49d9d63ae94ba044128e39b53ab50da00c08efde6ea7474ffca32fca"} Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.707755 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.707961 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.711794 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" event={"ID":"3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0","Type":"ContainerStarted","Data":"d51ba024d340b504e20579dee8976f7d30e24c521456480107df1d1d899dec33"} Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.712029 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.714706 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.714771 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4" event={"ID":"2da8d731-f76e-472f-bb05-b2242fb9a652","Type":"ContainerDied","Data":"29bfdbe49315178c37aed49f42ceab871fffb7a7d0b660f0317e5478b8f84a02"} Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.714867 4735 scope.go:117] "RemoveContainer" containerID="e377a3308baafe60f81aa8f63d7e6d2c6553997e36c10002f44b98cb04b05cb0" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.722166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" event={"ID":"a2ee43fb-95a1-40e2-837c-38f8e73c0357","Type":"ContainerStarted","Data":"42cac2f350a987fce84f2602f43e63e77740d21ba48f9c58948df0fa189254fc"} Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.726295 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.726837 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.727452 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" event={"ID":"e095b65e-6f13-4a89-8418-78f5bda82cb5","Type":"ContainerDied","Data":"3075a3c5776922c438d94b69f2819c1cf146fd543a7148c67342576a310e4eed"} Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.727632 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zqccv" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.743827 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.745252 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" podStartSLOduration=2.031294558 podStartE2EDuration="10.745227713s" podCreationTimestamp="2025-11-22 08:15:19 +0000 UTC" firstStartedPulling="2025-11-22 08:15:20.548655014 +0000 UTC m=+742.152993619" lastFinishedPulling="2025-11-22 08:15:29.262588169 +0000 UTC m=+750.866926774" observedRunningTime="2025-11-22 08:15:29.735522143 +0000 UTC m=+751.339860758" watchObservedRunningTime="2025-11-22 08:15:29.745227713 +0000 UTC m=+751.349566318" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.758987 4735 scope.go:117] "RemoveContainer" containerID="fccf5d6f15567cd849c87eedb7e31d04969c435541bbc3011e279c73ed013406" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.775191 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" podStartSLOduration=1.926050488 podStartE2EDuration="10.775167648s" podCreationTimestamp="2025-11-22 08:15:19 +0000 UTC" firstStartedPulling="2025-11-22 08:15:20.399215881 +0000 UTC m=+742.003554486" lastFinishedPulling="2025-11-22 08:15:29.248333031 +0000 UTC m=+750.852671646" observedRunningTime="2025-11-22 08:15:29.767855503 +0000 UTC m=+751.372194118" watchObservedRunningTime="2025-11-22 08:15:29.775167648 +0000 UTC m=+751.379506253" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.815058 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zqccv"] Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.824710 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zqccv"] Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.833523 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4"] Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.835708 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5rsm4"] Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.976576 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:29 crc kubenswrapper[4735]: I1122 08:15:29.990513 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" Nov 22 08:15:30 crc kubenswrapper[4735]: I1122 08:15:30.736698 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" event={"ID":"a2ee43fb-95a1-40e2-837c-38f8e73c0357","Type":"ContainerStarted","Data":"fec9bd054d13540b625bd83da7be23a373f9a8f2c9cefcb85c0d69db2dc9d099"} Nov 22 08:15:30 crc kubenswrapper[4735]: I1122 08:15:30.737019 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:30 crc kubenswrapper[4735]: I1122 08:15:30.743558 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" Nov 22 08:15:30 crc kubenswrapper[4735]: I1122 08:15:30.756227 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84bd655b67-d5jgl" podStartSLOduration=3.756212035 podStartE2EDuration="3.756212035s" podCreationTimestamp="2025-11-22 08:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:30.755170257 +0000 UTC m=+752.359508972" watchObservedRunningTime="2025-11-22 08:15:30.756212035 +0000 UTC m=+752.360550640" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.272263 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2da8d731-f76e-472f-bb05-b2242fb9a652" path="/var/lib/kubelet/pods/2da8d731-f76e-472f-bb05-b2242fb9a652/volumes" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.272989 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e095b65e-6f13-4a89-8418-78f5bda82cb5" path="/var/lib/kubelet/pods/e095b65e-6f13-4a89-8418-78f5bda82cb5/volumes" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.452594 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5cdfddb84-bms8j"] Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.453667 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.455290 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.455798 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.457387 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.460132 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.463808 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.465809 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.477642 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5cdfddb84-bms8j"] Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.484862 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.540021 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-client-ca\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.540071 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-config\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.540114 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-serving-cert\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.540145 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwb4k\" (UniqueName: \"kubernetes.io/projected/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-kube-api-access-rwb4k\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.541589 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-proxy-ca-bundles\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.642903 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-client-ca\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.642950 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-config\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.643002 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-serving-cert\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.643039 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwb4k\" (UniqueName: \"kubernetes.io/projected/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-kube-api-access-rwb4k\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.643076 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-proxy-ca-bundles\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.643766 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-client-ca\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.644179 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-proxy-ca-bundles\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.644725 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-config\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.649149 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-serving-cert\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.660513 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwb4k\" (UniqueName: \"kubernetes.io/projected/fced6d30-d46c-4ea7-80db-47a1d4a6cfca-kube-api-access-rwb4k\") pod \"controller-manager-5cdfddb84-bms8j\" (UID: \"fced6d30-d46c-4ea7-80db-47a1d4a6cfca\") " pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:31 crc kubenswrapper[4735]: I1122 08:15:31.790578 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:32 crc kubenswrapper[4735]: I1122 08:15:32.214728 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5cdfddb84-bms8j"] Nov 22 08:15:32 crc kubenswrapper[4735]: I1122 08:15:32.757663 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" event={"ID":"fced6d30-d46c-4ea7-80db-47a1d4a6cfca","Type":"ContainerStarted","Data":"8c36c66a0cf0c7ea8c7bd50aa2e54e884e7581207c11eb4197770c05f443e8de"} Nov 22 08:15:32 crc kubenswrapper[4735]: I1122 08:15:32.757939 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:32 crc kubenswrapper[4735]: I1122 08:15:32.757951 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" event={"ID":"fced6d30-d46c-4ea7-80db-47a1d4a6cfca","Type":"ContainerStarted","Data":"6893a70742267696435a86fcd789ba2cd758f851fdc367936e48fbf203ca30ee"} Nov 22 08:15:32 crc kubenswrapper[4735]: I1122 08:15:32.763141 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" Nov 22 08:15:32 crc kubenswrapper[4735]: I1122 08:15:32.774163 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5cdfddb84-bms8j" podStartSLOduration=5.774142998 podStartE2EDuration="5.774142998s" podCreationTimestamp="2025-11-22 08:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:32.772666147 +0000 UTC m=+754.377004782" watchObservedRunningTime="2025-11-22 08:15:32.774142998 +0000 UTC m=+754.378481613" Nov 22 08:15:35 crc kubenswrapper[4735]: I1122 08:15:35.662696 4735 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 08:15:38 crc kubenswrapper[4735]: I1122 08:15:38.957384 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hvxhx" Nov 22 08:15:39 crc kubenswrapper[4735]: I1122 08:15:39.130627 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-zhskj" Nov 22 08:15:39 crc kubenswrapper[4735]: I1122 08:15:39.282479 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-hxcvg" Nov 22 08:15:40 crc kubenswrapper[4735]: I1122 08:15:40.260007 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Nov 22 08:15:40 crc kubenswrapper[4735]: I1122 08:15:40.431064 4735 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 22 08:15:40 crc kubenswrapper[4735]: I1122 08:15:40.431165 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="d3b1eee2-6b00-45b3-95bd-9e3005004633" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 22 08:15:40 crc kubenswrapper[4735]: I1122 08:15:40.452515 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.132081 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.132840 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.132908 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.133948 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a3aab057b98bc5c57ec42d5f2111ef56b0a63d358e465d259bfc8b9b235ad36f"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.134044 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://a3aab057b98bc5c57ec42d5f2111ef56b0a63d358e465d259bfc8b9b235ad36f" gracePeriod=600 Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.865450 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="a3aab057b98bc5c57ec42d5f2111ef56b0a63d358e465d259bfc8b9b235ad36f" exitCode=0 Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.865540 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"a3aab057b98bc5c57ec42d5f2111ef56b0a63d358e465d259bfc8b9b235ad36f"} Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.866142 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"8ac3677e0ebb55329e2a912a3a81033db8ba05ef418621a9d574aafba4dff19d"} Nov 22 08:15:46 crc kubenswrapper[4735]: I1122 08:15:46.866172 4735 scope.go:117] "RemoveContainer" containerID="f7aa357304ffacda92a92578afeb23951ae04837d67816d3ca4f766ea770f256" Nov 22 08:15:50 crc kubenswrapper[4735]: I1122 08:15:50.428001 4735 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 22 08:15:50 crc kubenswrapper[4735]: I1122 08:15:50.429194 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="d3b1eee2-6b00-45b3-95bd-9e3005004633" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 22 08:16:00 crc kubenswrapper[4735]: I1122 08:16:00.432203 4735 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 22 08:16:00 crc kubenswrapper[4735]: I1122 08:16:00.432629 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="d3b1eee2-6b00-45b3-95bd-9e3005004633" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 22 08:16:10 crc kubenswrapper[4735]: I1122 08:16:10.428777 4735 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 22 08:16:10 crc kubenswrapper[4735]: I1122 08:16:10.429544 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="d3b1eee2-6b00-45b3-95bd-9e3005004633" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.011499 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v7f2r"] Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.013887 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.021625 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-utilities\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.021786 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smbmr\" (UniqueName: \"kubernetes.io/projected/810c15f7-b37c-41a8-bdc9-afab2d886c27-kube-api-access-smbmr\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.021944 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-catalog-content\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.036141 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v7f2r"] Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.123266 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smbmr\" (UniqueName: \"kubernetes.io/projected/810c15f7-b37c-41a8-bdc9-afab2d886c27-kube-api-access-smbmr\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.123368 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-catalog-content\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.123401 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-utilities\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.124072 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-catalog-content\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.124246 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-utilities\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.148598 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smbmr\" (UniqueName: \"kubernetes.io/projected/810c15f7-b37c-41a8-bdc9-afab2d886c27-kube-api-access-smbmr\") pod \"certified-operators-v7f2r\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.346939 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:13 crc kubenswrapper[4735]: I1122 08:16:13.808687 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v7f2r"] Nov 22 08:16:14 crc kubenswrapper[4735]: I1122 08:16:14.083022 4735 generic.go:334] "Generic (PLEG): container finished" podID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerID="6a0b4dbdf44d1bf394d6cf912e37b608c0d0934c99a74892c2c081ea50e93063" exitCode=0 Nov 22 08:16:14 crc kubenswrapper[4735]: I1122 08:16:14.083079 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7f2r" event={"ID":"810c15f7-b37c-41a8-bdc9-afab2d886c27","Type":"ContainerDied","Data":"6a0b4dbdf44d1bf394d6cf912e37b608c0d0934c99a74892c2c081ea50e93063"} Nov 22 08:16:14 crc kubenswrapper[4735]: I1122 08:16:14.083289 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7f2r" event={"ID":"810c15f7-b37c-41a8-bdc9-afab2d886c27","Type":"ContainerStarted","Data":"f532cabd4f6e1ccfeaad6e21e26a7e6c51ff4facd59f34e7b81ee5aed603b6fd"} Nov 22 08:16:15 crc kubenswrapper[4735]: I1122 08:16:15.093960 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7f2r" event={"ID":"810c15f7-b37c-41a8-bdc9-afab2d886c27","Type":"ContainerStarted","Data":"a4b70a23c32502548fd144717416998e479d6ec905bbb7c84e154630380bbf94"} Nov 22 08:16:16 crc kubenswrapper[4735]: I1122 08:16:16.103670 4735 generic.go:334] "Generic (PLEG): container finished" podID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerID="a4b70a23c32502548fd144717416998e479d6ec905bbb7c84e154630380bbf94" exitCode=0 Nov 22 08:16:16 crc kubenswrapper[4735]: I1122 08:16:16.103898 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7f2r" event={"ID":"810c15f7-b37c-41a8-bdc9-afab2d886c27","Type":"ContainerDied","Data":"a4b70a23c32502548fd144717416998e479d6ec905bbb7c84e154630380bbf94"} Nov 22 08:16:16 crc kubenswrapper[4735]: I1122 08:16:16.103928 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7f2r" event={"ID":"810c15f7-b37c-41a8-bdc9-afab2d886c27","Type":"ContainerStarted","Data":"78cc7e7b5c23be3206eae92b566f58cc18d3d7ee4b848db9ba3a99b8f7e30c77"} Nov 22 08:16:16 crc kubenswrapper[4735]: I1122 08:16:16.121451 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v7f2r" podStartSLOduration=2.630095884 podStartE2EDuration="4.121434205s" podCreationTimestamp="2025-11-22 08:16:12 +0000 UTC" firstStartedPulling="2025-11-22 08:16:14.084789812 +0000 UTC m=+795.689128427" lastFinishedPulling="2025-11-22 08:16:15.576128143 +0000 UTC m=+797.180466748" observedRunningTime="2025-11-22 08:16:16.119268396 +0000 UTC m=+797.723607021" watchObservedRunningTime="2025-11-22 08:16:16.121434205 +0000 UTC m=+797.725772810" Nov 22 08:16:20 crc kubenswrapper[4735]: I1122 08:16:20.431650 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Nov 22 08:16:23 crc kubenswrapper[4735]: I1122 08:16:23.347633 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:23 crc kubenswrapper[4735]: I1122 08:16:23.348345 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:23 crc kubenswrapper[4735]: I1122 08:16:23.420928 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:24 crc kubenswrapper[4735]: I1122 08:16:24.222095 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:24 crc kubenswrapper[4735]: I1122 08:16:24.281802 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v7f2r"] Nov 22 08:16:26 crc kubenswrapper[4735]: I1122 08:16:26.181589 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v7f2r" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerName="registry-server" containerID="cri-o://78cc7e7b5c23be3206eae92b566f58cc18d3d7ee4b848db9ba3a99b8f7e30c77" gracePeriod=2 Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.188757 4735 generic.go:334] "Generic (PLEG): container finished" podID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerID="78cc7e7b5c23be3206eae92b566f58cc18d3d7ee4b848db9ba3a99b8f7e30c77" exitCode=0 Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.188791 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7f2r" event={"ID":"810c15f7-b37c-41a8-bdc9-afab2d886c27","Type":"ContainerDied","Data":"78cc7e7b5c23be3206eae92b566f58cc18d3d7ee4b848db9ba3a99b8f7e30c77"} Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.751124 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.872470 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smbmr\" (UniqueName: \"kubernetes.io/projected/810c15f7-b37c-41a8-bdc9-afab2d886c27-kube-api-access-smbmr\") pod \"810c15f7-b37c-41a8-bdc9-afab2d886c27\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.872566 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-utilities\") pod \"810c15f7-b37c-41a8-bdc9-afab2d886c27\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.872624 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-catalog-content\") pod \"810c15f7-b37c-41a8-bdc9-afab2d886c27\" (UID: \"810c15f7-b37c-41a8-bdc9-afab2d886c27\") " Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.873658 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-utilities" (OuterVolumeSpecName: "utilities") pod "810c15f7-b37c-41a8-bdc9-afab2d886c27" (UID: "810c15f7-b37c-41a8-bdc9-afab2d886c27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.878902 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/810c15f7-b37c-41a8-bdc9-afab2d886c27-kube-api-access-smbmr" (OuterVolumeSpecName: "kube-api-access-smbmr") pod "810c15f7-b37c-41a8-bdc9-afab2d886c27" (UID: "810c15f7-b37c-41a8-bdc9-afab2d886c27"). InnerVolumeSpecName "kube-api-access-smbmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.917942 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "810c15f7-b37c-41a8-bdc9-afab2d886c27" (UID: "810c15f7-b37c-41a8-bdc9-afab2d886c27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.975183 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smbmr\" (UniqueName: \"kubernetes.io/projected/810c15f7-b37c-41a8-bdc9-afab2d886c27-kube-api-access-smbmr\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.975230 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:27 crc kubenswrapper[4735]: I1122 08:16:27.975245 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810c15f7-b37c-41a8-bdc9-afab2d886c27-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:28 crc kubenswrapper[4735]: I1122 08:16:28.198612 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7f2r" event={"ID":"810c15f7-b37c-41a8-bdc9-afab2d886c27","Type":"ContainerDied","Data":"f532cabd4f6e1ccfeaad6e21e26a7e6c51ff4facd59f34e7b81ee5aed603b6fd"} Nov 22 08:16:28 crc kubenswrapper[4735]: I1122 08:16:28.198661 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7f2r" Nov 22 08:16:28 crc kubenswrapper[4735]: I1122 08:16:28.198691 4735 scope.go:117] "RemoveContainer" containerID="78cc7e7b5c23be3206eae92b566f58cc18d3d7ee4b848db9ba3a99b8f7e30c77" Nov 22 08:16:28 crc kubenswrapper[4735]: I1122 08:16:28.215484 4735 scope.go:117] "RemoveContainer" containerID="a4b70a23c32502548fd144717416998e479d6ec905bbb7c84e154630380bbf94" Nov 22 08:16:28 crc kubenswrapper[4735]: I1122 08:16:28.235794 4735 scope.go:117] "RemoveContainer" containerID="6a0b4dbdf44d1bf394d6cf912e37b608c0d0934c99a74892c2c081ea50e93063" Nov 22 08:16:28 crc kubenswrapper[4735]: I1122 08:16:28.240519 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v7f2r"] Nov 22 08:16:28 crc kubenswrapper[4735]: I1122 08:16:28.246147 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v7f2r"] Nov 22 08:16:29 crc kubenswrapper[4735]: I1122 08:16:29.271771 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" path="/var/lib/kubelet/pods/810c15f7-b37c-41a8-bdc9-afab2d886c27/volumes" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.511765 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zp7pl"] Nov 22 08:16:34 crc kubenswrapper[4735]: E1122 08:16:34.512728 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerName="extract-utilities" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.512770 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerName="extract-utilities" Nov 22 08:16:34 crc kubenswrapper[4735]: E1122 08:16:34.512790 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerName="extract-content" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.512796 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerName="extract-content" Nov 22 08:16:34 crc kubenswrapper[4735]: E1122 08:16:34.512805 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerName="registry-server" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.512811 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerName="registry-server" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.512961 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="810c15f7-b37c-41a8-bdc9-afab2d886c27" containerName="registry-server" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.513973 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.524784 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zp7pl"] Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.688667 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf42z\" (UniqueName: \"kubernetes.io/projected/d3a55ad6-50dc-40da-8a55-127d2c167c3a-kube-api-access-vf42z\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.688765 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-utilities\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.688833 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-catalog-content\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.790877 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf42z\" (UniqueName: \"kubernetes.io/projected/d3a55ad6-50dc-40da-8a55-127d2c167c3a-kube-api-access-vf42z\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.791272 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-utilities\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.791298 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-catalog-content\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.791701 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-utilities\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.791723 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-catalog-content\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.816221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf42z\" (UniqueName: \"kubernetes.io/projected/d3a55ad6-50dc-40da-8a55-127d2c167c3a-kube-api-access-vf42z\") pod \"redhat-marketplace-zp7pl\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:34 crc kubenswrapper[4735]: I1122 08:16:34.829838 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:35 crc kubenswrapper[4735]: I1122 08:16:35.249441 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zp7pl"] Nov 22 08:16:36 crc kubenswrapper[4735]: I1122 08:16:36.263866 4735 generic.go:334] "Generic (PLEG): container finished" podID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerID="d2e6afcf313f61bc19b8216021057feb369f29ed462e882b1a2746664347b465" exitCode=0 Nov 22 08:16:36 crc kubenswrapper[4735]: I1122 08:16:36.263942 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zp7pl" event={"ID":"d3a55ad6-50dc-40da-8a55-127d2c167c3a","Type":"ContainerDied","Data":"d2e6afcf313f61bc19b8216021057feb369f29ed462e882b1a2746664347b465"} Nov 22 08:16:36 crc kubenswrapper[4735]: I1122 08:16:36.264241 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zp7pl" event={"ID":"d3a55ad6-50dc-40da-8a55-127d2c167c3a","Type":"ContainerStarted","Data":"e9eb074fc64765dfdc29b2e15eafefb6716660a178e8799815acbe7522431e46"} Nov 22 08:16:37 crc kubenswrapper[4735]: I1122 08:16:37.294654 4735 generic.go:334] "Generic (PLEG): container finished" podID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerID="8a3ff17be5787ee209e8b32e04b246ee49aa283a717ea8d0d0364755b9696e8f" exitCode=0 Nov 22 08:16:37 crc kubenswrapper[4735]: I1122 08:16:37.294762 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zp7pl" event={"ID":"d3a55ad6-50dc-40da-8a55-127d2c167c3a","Type":"ContainerDied","Data":"8a3ff17be5787ee209e8b32e04b246ee49aa283a717ea8d0d0364755b9696e8f"} Nov 22 08:16:38 crc kubenswrapper[4735]: I1122 08:16:38.304968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zp7pl" event={"ID":"d3a55ad6-50dc-40da-8a55-127d2c167c3a","Type":"ContainerStarted","Data":"99f21919b35e203374ffdd5081af7d6d4119409529dc09e06ecaa4f71e4358a1"} Nov 22 08:16:38 crc kubenswrapper[4735]: I1122 08:16:38.330684 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zp7pl" podStartSLOduration=2.890323892 podStartE2EDuration="4.330664054s" podCreationTimestamp="2025-11-22 08:16:34 +0000 UTC" firstStartedPulling="2025-11-22 08:16:36.270368741 +0000 UTC m=+817.874707356" lastFinishedPulling="2025-11-22 08:16:37.710708873 +0000 UTC m=+819.315047518" observedRunningTime="2025-11-22 08:16:38.326027168 +0000 UTC m=+819.930365793" watchObservedRunningTime="2025-11-22 08:16:38.330664054 +0000 UTC m=+819.935002659" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.770745 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-zsks9"] Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.774617 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.779952 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.781892 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.782129 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.782354 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.782636 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-pjqfz" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.786835 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.788619 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-zsks9"] Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874014 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-sa-token\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874105 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-entrypoint\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874160 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874203 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/223bc6a4-4dbc-4084-b56e-525d52d1afb7-datadir\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874275 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config-openshift-service-cacrt\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874340 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-trusted-ca\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874391 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874433 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/223bc6a4-4dbc-4084-b56e-525d52d1afb7-tmp\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874620 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd629\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-kube-api-access-qd629\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874726 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.874760 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-token\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.933874 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-zsks9"] Nov 22 08:16:39 crc kubenswrapper[4735]: E1122 08:16:39.934727 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-qd629 metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-zsks9" podUID="223bc6a4-4dbc-4084-b56e-525d52d1afb7" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976264 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd629\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-kube-api-access-qd629\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976372 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976403 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-token\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976443 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-sa-token\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976475 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-entrypoint\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976524 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976550 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/223bc6a4-4dbc-4084-b56e-525d52d1afb7-datadir\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976605 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config-openshift-service-cacrt\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976639 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-trusted-ca\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976665 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.976699 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/223bc6a4-4dbc-4084-b56e-525d52d1afb7-tmp\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: E1122 08:16:39.978342 4735 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Nov 22 08:16:39 crc kubenswrapper[4735]: E1122 08:16:39.978459 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics podName:223bc6a4-4dbc-4084-b56e-525d52d1afb7 nodeName:}" failed. No retries permitted until 2025-11-22 08:16:40.478426007 +0000 UTC m=+822.082764612 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics") pod "collector-zsks9" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7") : secret "collector-metrics" not found Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.978698 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config-openshift-service-cacrt\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.978708 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-entrypoint\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: E1122 08:16:39.978858 4735 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Nov 22 08:16:39 crc kubenswrapper[4735]: E1122 08:16:39.978923 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver podName:223bc6a4-4dbc-4084-b56e-525d52d1afb7 nodeName:}" failed. No retries permitted until 2025-11-22 08:16:40.47890226 +0000 UTC m=+822.083240865 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver") pod "collector-zsks9" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7") : secret "collector-syslog-receiver" not found Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.978953 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/223bc6a4-4dbc-4084-b56e-525d52d1afb7-datadir\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.979832 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.979880 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-trusted-ca\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.983652 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/223bc6a4-4dbc-4084-b56e-525d52d1afb7-tmp\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.984354 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-token\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:39 crc kubenswrapper[4735]: I1122 08:16:39.997871 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd629\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-kube-api-access-qd629\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.000696 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-sa-token\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.318598 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-zsks9" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.328759 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-zsks9" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.381870 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.382152 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/223bc6a4-4dbc-4084-b56e-525d52d1afb7-datadir\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.382219 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/223bc6a4-4dbc-4084-b56e-525d52d1afb7-datadir" (OuterVolumeSpecName: "datadir") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.382337 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-sa-token\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.382406 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config" (OuterVolumeSpecName: "config") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.382562 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/223bc6a4-4dbc-4084-b56e-525d52d1afb7-tmp\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.382740 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-trusted-ca\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.382854 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-entrypoint\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.383095 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-token\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.383194 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd629\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-kube-api-access-qd629\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.383278 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config-openshift-service-cacrt\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.384058 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.384212 4735 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/223bc6a4-4dbc-4084-b56e-525d52d1afb7-datadir\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.383323 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.383740 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.384241 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.386843 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-token" (OuterVolumeSpecName: "collector-token") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.387620 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/223bc6a4-4dbc-4084-b56e-525d52d1afb7-tmp" (OuterVolumeSpecName: "tmp") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.388507 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-kube-api-access-qd629" (OuterVolumeSpecName: "kube-api-access-qd629") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "kube-api-access-qd629". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.388553 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-sa-token" (OuterVolumeSpecName: "sa-token") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486225 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486285 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486365 4735 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/223bc6a4-4dbc-4084-b56e-525d52d1afb7-tmp\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486377 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486389 4735 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-entrypoint\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486399 4735 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-token\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486408 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd629\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-kube-api-access-qd629\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486418 4735 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/223bc6a4-4dbc-4084-b56e-525d52d1afb7-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.486428 4735 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/223bc6a4-4dbc-4084-b56e-525d52d1afb7-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.490246 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.491617 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics\") pod \"collector-zsks9\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " pod="openshift-logging/collector-zsks9" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.587717 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.588378 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics\") pod \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\" (UID: \"223bc6a4-4dbc-4084-b56e-525d52d1afb7\") " Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.592378 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics" (OuterVolumeSpecName: "metrics") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.592766 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "223bc6a4-4dbc-4084-b56e-525d52d1afb7" (UID: "223bc6a4-4dbc-4084-b56e-525d52d1afb7"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.691898 4735 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:40 crc kubenswrapper[4735]: I1122 08:16:40.691988 4735 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/223bc6a4-4dbc-4084-b56e-525d52d1afb7-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.329251 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-zsks9" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.406586 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-zsks9"] Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.420237 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-zsks9"] Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.427382 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-dmrkn"] Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.429540 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.432292 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.432589 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.432928 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-pjqfz" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.432969 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-dmrkn"] Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.437501 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.438584 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.441876 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.505812 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4ea5da34-9c92-4c02-a674-54618b1c5888-tmp\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.505891 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/4ea5da34-9c92-4c02-a674-54618b1c5888-sa-token\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.505955 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-collector-syslog-receiver\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.505996 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/4ea5da34-9c92-4c02-a674-54618b1c5888-datadir\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.506027 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-metrics\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.506066 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-trusted-ca\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.506121 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-config-openshift-service-cacrt\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.506173 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-config\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.506244 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxv5q\" (UniqueName: \"kubernetes.io/projected/4ea5da34-9c92-4c02-a674-54618b1c5888-kube-api-access-zxv5q\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.506287 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-collector-token\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.506339 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-entrypoint\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.607757 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-entrypoint\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.607873 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4ea5da34-9c92-4c02-a674-54618b1c5888-tmp\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.607908 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/4ea5da34-9c92-4c02-a674-54618b1c5888-sa-token\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.607951 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-collector-syslog-receiver\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.607977 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/4ea5da34-9c92-4c02-a674-54618b1c5888-datadir\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.607997 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-metrics\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.608027 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-trusted-ca\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.608053 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-config-openshift-service-cacrt\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.608085 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-config\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.608136 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxv5q\" (UniqueName: \"kubernetes.io/projected/4ea5da34-9c92-4c02-a674-54618b1c5888-kube-api-access-zxv5q\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.608167 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-collector-token\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.608711 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/4ea5da34-9c92-4c02-a674-54618b1c5888-datadir\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.609334 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-entrypoint\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.609557 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-config-openshift-service-cacrt\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.609745 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-trusted-ca\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.611072 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea5da34-9c92-4c02-a674-54618b1c5888-config\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.612406 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4ea5da34-9c92-4c02-a674-54618b1c5888-tmp\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.613226 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-collector-token\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.613719 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-collector-syslog-receiver\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.620602 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/4ea5da34-9c92-4c02-a674-54618b1c5888-metrics\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.634513 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/4ea5da34-9c92-4c02-a674-54618b1c5888-sa-token\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.641781 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxv5q\" (UniqueName: \"kubernetes.io/projected/4ea5da34-9c92-4c02-a674-54618b1c5888-kube-api-access-zxv5q\") pod \"collector-dmrkn\" (UID: \"4ea5da34-9c92-4c02-a674-54618b1c5888\") " pod="openshift-logging/collector-dmrkn" Nov 22 08:16:41 crc kubenswrapper[4735]: I1122 08:16:41.748076 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-dmrkn" Nov 22 08:16:42 crc kubenswrapper[4735]: I1122 08:16:42.194808 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-dmrkn"] Nov 22 08:16:42 crc kubenswrapper[4735]: I1122 08:16:42.338605 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-dmrkn" event={"ID":"4ea5da34-9c92-4c02-a674-54618b1c5888","Type":"ContainerStarted","Data":"4655a59299358693ed8317a9ca90645c44d40329a808376b7aa4de9d089f0301"} Nov 22 08:16:43 crc kubenswrapper[4735]: I1122 08:16:43.273427 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="223bc6a4-4dbc-4084-b56e-525d52d1afb7" path="/var/lib/kubelet/pods/223bc6a4-4dbc-4084-b56e-525d52d1afb7/volumes" Nov 22 08:16:44 crc kubenswrapper[4735]: I1122 08:16:44.830660 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:44 crc kubenswrapper[4735]: I1122 08:16:44.830988 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:44 crc kubenswrapper[4735]: I1122 08:16:44.885856 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:45 crc kubenswrapper[4735]: I1122 08:16:45.411903 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:45 crc kubenswrapper[4735]: I1122 08:16:45.457939 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zp7pl"] Nov 22 08:16:47 crc kubenswrapper[4735]: I1122 08:16:47.382094 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zp7pl" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerName="registry-server" containerID="cri-o://99f21919b35e203374ffdd5081af7d6d4119409529dc09e06ecaa4f71e4358a1" gracePeriod=2 Nov 22 08:16:47 crc kubenswrapper[4735]: I1122 08:16:47.976340 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t5vpq"] Nov 22 08:16:47 crc kubenswrapper[4735]: I1122 08:16:47.980284 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:47 crc kubenswrapper[4735]: I1122 08:16:47.990489 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t5vpq"] Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.008077 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-catalog-content\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.008134 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-224dx\" (UniqueName: \"kubernetes.io/projected/a7e069fd-4e58-471d-bc39-4c759edd9c9f-kube-api-access-224dx\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.008404 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-utilities\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.109990 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-utilities\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.110189 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-catalog-content\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.110219 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-224dx\" (UniqueName: \"kubernetes.io/projected/a7e069fd-4e58-471d-bc39-4c759edd9c9f-kube-api-access-224dx\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.115127 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-catalog-content\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.115972 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-utilities\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.142445 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-224dx\" (UniqueName: \"kubernetes.io/projected/a7e069fd-4e58-471d-bc39-4c759edd9c9f-kube-api-access-224dx\") pod \"community-operators-t5vpq\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.330174 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.393705 4735 generic.go:334] "Generic (PLEG): container finished" podID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerID="99f21919b35e203374ffdd5081af7d6d4119409529dc09e06ecaa4f71e4358a1" exitCode=0 Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.393759 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zp7pl" event={"ID":"d3a55ad6-50dc-40da-8a55-127d2c167c3a","Type":"ContainerDied","Data":"99f21919b35e203374ffdd5081af7d6d4119409529dc09e06ecaa4f71e4358a1"} Nov 22 08:16:48 crc kubenswrapper[4735]: I1122 08:16:48.859861 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.023178 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf42z\" (UniqueName: \"kubernetes.io/projected/d3a55ad6-50dc-40da-8a55-127d2c167c3a-kube-api-access-vf42z\") pod \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.023306 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-catalog-content\") pod \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.023460 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-utilities\") pod \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\" (UID: \"d3a55ad6-50dc-40da-8a55-127d2c167c3a\") " Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.027779 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a55ad6-50dc-40da-8a55-127d2c167c3a-kube-api-access-vf42z" (OuterVolumeSpecName: "kube-api-access-vf42z") pod "d3a55ad6-50dc-40da-8a55-127d2c167c3a" (UID: "d3a55ad6-50dc-40da-8a55-127d2c167c3a"). InnerVolumeSpecName "kube-api-access-vf42z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.028349 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-utilities" (OuterVolumeSpecName: "utilities") pod "d3a55ad6-50dc-40da-8a55-127d2c167c3a" (UID: "d3a55ad6-50dc-40da-8a55-127d2c167c3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.042918 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t5vpq"] Nov 22 08:16:49 crc kubenswrapper[4735]: W1122 08:16:49.044921 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7e069fd_4e58_471d_bc39_4c759edd9c9f.slice/crio-5982ee62be6217ab5349ac1daa041cefd59671e184a475c87a9e0669b316ca39 WatchSource:0}: Error finding container 5982ee62be6217ab5349ac1daa041cefd59671e184a475c87a9e0669b316ca39: Status 404 returned error can't find the container with id 5982ee62be6217ab5349ac1daa041cefd59671e184a475c87a9e0669b316ca39 Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.059103 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3a55ad6-50dc-40da-8a55-127d2c167c3a" (UID: "d3a55ad6-50dc-40da-8a55-127d2c167c3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.124511 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.124997 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf42z\" (UniqueName: \"kubernetes.io/projected/d3a55ad6-50dc-40da-8a55-127d2c167c3a-kube-api-access-vf42z\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.125084 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a55ad6-50dc-40da-8a55-127d2c167c3a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.410998 4735 generic.go:334] "Generic (PLEG): container finished" podID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerID="edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af" exitCode=0 Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.411070 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5vpq" event={"ID":"a7e069fd-4e58-471d-bc39-4c759edd9c9f","Type":"ContainerDied","Data":"edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af"} Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.411097 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5vpq" event={"ID":"a7e069fd-4e58-471d-bc39-4c759edd9c9f","Type":"ContainerStarted","Data":"5982ee62be6217ab5349ac1daa041cefd59671e184a475c87a9e0669b316ca39"} Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.413340 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zp7pl" event={"ID":"d3a55ad6-50dc-40da-8a55-127d2c167c3a","Type":"ContainerDied","Data":"e9eb074fc64765dfdc29b2e15eafefb6716660a178e8799815acbe7522431e46"} Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.413425 4735 scope.go:117] "RemoveContainer" containerID="99f21919b35e203374ffdd5081af7d6d4119409529dc09e06ecaa4f71e4358a1" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.413360 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zp7pl" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.419894 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-dmrkn" event={"ID":"4ea5da34-9c92-4c02-a674-54618b1c5888","Type":"ContainerStarted","Data":"d084371f2aa59c2fab515e0e56220923474c279f80c51f5afe676e837df2d9ac"} Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.445536 4735 scope.go:117] "RemoveContainer" containerID="8a3ff17be5787ee209e8b32e04b246ee49aa283a717ea8d0d0364755b9696e8f" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.446718 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zp7pl"] Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.458063 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zp7pl"] Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.465986 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-dmrkn" podStartSLOduration=1.9951077910000001 podStartE2EDuration="8.465964933s" podCreationTimestamp="2025-11-22 08:16:41 +0000 UTC" firstStartedPulling="2025-11-22 08:16:42.204905676 +0000 UTC m=+823.809244291" lastFinishedPulling="2025-11-22 08:16:48.675762828 +0000 UTC m=+830.280101433" observedRunningTime="2025-11-22 08:16:49.459626531 +0000 UTC m=+831.063965136" watchObservedRunningTime="2025-11-22 08:16:49.465964933 +0000 UTC m=+831.070303578" Nov 22 08:16:49 crc kubenswrapper[4735]: I1122 08:16:49.482769 4735 scope.go:117] "RemoveContainer" containerID="d2e6afcf313f61bc19b8216021057feb369f29ed462e882b1a2746664347b465" Nov 22 08:16:51 crc kubenswrapper[4735]: I1122 08:16:51.278652 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" path="/var/lib/kubelet/pods/d3a55ad6-50dc-40da-8a55-127d2c167c3a/volumes" Nov 22 08:16:51 crc kubenswrapper[4735]: I1122 08:16:51.435530 4735 generic.go:334] "Generic (PLEG): container finished" podID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerID="849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2" exitCode=0 Nov 22 08:16:51 crc kubenswrapper[4735]: I1122 08:16:51.435606 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5vpq" event={"ID":"a7e069fd-4e58-471d-bc39-4c759edd9c9f","Type":"ContainerDied","Data":"849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2"} Nov 22 08:16:52 crc kubenswrapper[4735]: I1122 08:16:52.445540 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5vpq" event={"ID":"a7e069fd-4e58-471d-bc39-4c759edd9c9f","Type":"ContainerStarted","Data":"76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6"} Nov 22 08:16:52 crc kubenswrapper[4735]: I1122 08:16:52.477107 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t5vpq" podStartSLOduration=3.06252455 podStartE2EDuration="5.477086886s" podCreationTimestamp="2025-11-22 08:16:47 +0000 UTC" firstStartedPulling="2025-11-22 08:16:49.414555262 +0000 UTC m=+831.018893867" lastFinishedPulling="2025-11-22 08:16:51.829117598 +0000 UTC m=+833.433456203" observedRunningTime="2025-11-22 08:16:52.469044529 +0000 UTC m=+834.073383164" watchObservedRunningTime="2025-11-22 08:16:52.477086886 +0000 UTC m=+834.081425501" Nov 22 08:16:58 crc kubenswrapper[4735]: I1122 08:16:58.330608 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:58 crc kubenswrapper[4735]: I1122 08:16:58.331155 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:58 crc kubenswrapper[4735]: I1122 08:16:58.385529 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:58 crc kubenswrapper[4735]: I1122 08:16:58.582834 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:16:58 crc kubenswrapper[4735]: I1122 08:16:58.633151 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t5vpq"] Nov 22 08:17:00 crc kubenswrapper[4735]: I1122 08:17:00.505627 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t5vpq" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerName="registry-server" containerID="cri-o://76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6" gracePeriod=2 Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.039752 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6pjm5"] Nov 22 08:17:01 crc kubenswrapper[4735]: E1122 08:17:01.040555 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerName="registry-server" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.040583 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerName="registry-server" Nov 22 08:17:01 crc kubenswrapper[4735]: E1122 08:17:01.040604 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerName="extract-content" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.040613 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerName="extract-content" Nov 22 08:17:01 crc kubenswrapper[4735]: E1122 08:17:01.040643 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerName="extract-utilities" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.040654 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerName="extract-utilities" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.040861 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a55ad6-50dc-40da-8a55-127d2c167c3a" containerName="registry-server" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.042553 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.048952 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6pjm5"] Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.072799 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.143897 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-utilities\") pod \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.143990 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-224dx\" (UniqueName: \"kubernetes.io/projected/a7e069fd-4e58-471d-bc39-4c759edd9c9f-kube-api-access-224dx\") pod \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.144045 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-catalog-content\") pod \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\" (UID: \"a7e069fd-4e58-471d-bc39-4c759edd9c9f\") " Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.144222 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt2q7\" (UniqueName: \"kubernetes.io/projected/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-kube-api-access-qt2q7\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.144258 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-catalog-content\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.144444 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-utilities\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.144936 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-utilities" (OuterVolumeSpecName: "utilities") pod "a7e069fd-4e58-471d-bc39-4c759edd9c9f" (UID: "a7e069fd-4e58-471d-bc39-4c759edd9c9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.149235 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e069fd-4e58-471d-bc39-4c759edd9c9f-kube-api-access-224dx" (OuterVolumeSpecName: "kube-api-access-224dx") pod "a7e069fd-4e58-471d-bc39-4c759edd9c9f" (UID: "a7e069fd-4e58-471d-bc39-4c759edd9c9f"). InnerVolumeSpecName "kube-api-access-224dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.245070 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt2q7\" (UniqueName: \"kubernetes.io/projected/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-kube-api-access-qt2q7\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.245317 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-catalog-content\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.245365 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-utilities\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.245438 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.245450 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-224dx\" (UniqueName: \"kubernetes.io/projected/a7e069fd-4e58-471d-bc39-4c759edd9c9f-kube-api-access-224dx\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.245858 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-utilities\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.245993 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-catalog-content\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.262696 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt2q7\" (UniqueName: \"kubernetes.io/projected/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-kube-api-access-qt2q7\") pod \"redhat-operators-6pjm5\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.385297 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.519190 4735 generic.go:334] "Generic (PLEG): container finished" podID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerID="76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6" exitCode=0 Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.519240 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5vpq" event={"ID":"a7e069fd-4e58-471d-bc39-4c759edd9c9f","Type":"ContainerDied","Data":"76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6"} Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.519271 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5vpq" event={"ID":"a7e069fd-4e58-471d-bc39-4c759edd9c9f","Type":"ContainerDied","Data":"5982ee62be6217ab5349ac1daa041cefd59671e184a475c87a9e0669b316ca39"} Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.519301 4735 scope.go:117] "RemoveContainer" containerID="76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.519344 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t5vpq" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.551697 4735 scope.go:117] "RemoveContainer" containerID="849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.579766 4735 scope.go:117] "RemoveContainer" containerID="edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.600638 4735 scope.go:117] "RemoveContainer" containerID="76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6" Nov 22 08:17:01 crc kubenswrapper[4735]: E1122 08:17:01.601163 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6\": container with ID starting with 76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6 not found: ID does not exist" containerID="76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.601210 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6"} err="failed to get container status \"76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6\": rpc error: code = NotFound desc = could not find container \"76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6\": container with ID starting with 76c318dc9665078f5d3864bb9878b2e9b09a94037dba13180d085a06b4cd63a6 not found: ID does not exist" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.601235 4735 scope.go:117] "RemoveContainer" containerID="849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2" Nov 22 08:17:01 crc kubenswrapper[4735]: E1122 08:17:01.603445 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2\": container with ID starting with 849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2 not found: ID does not exist" containerID="849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.603489 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2"} err="failed to get container status \"849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2\": rpc error: code = NotFound desc = could not find container \"849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2\": container with ID starting with 849f81aeb0a76d60670e424e3259b0f19d502795c8b1e4699b9c9e844add80d2 not found: ID does not exist" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.603508 4735 scope.go:117] "RemoveContainer" containerID="edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af" Nov 22 08:17:01 crc kubenswrapper[4735]: E1122 08:17:01.603889 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af\": container with ID starting with edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af not found: ID does not exist" containerID="edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.603949 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af"} err="failed to get container status \"edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af\": rpc error: code = NotFound desc = could not find container \"edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af\": container with ID starting with edf82ff24e599e116548cf4bb669515ff370b6699cc4daa99d1b0a1f2dce53af not found: ID does not exist" Nov 22 08:17:01 crc kubenswrapper[4735]: I1122 08:17:01.836353 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6pjm5"] Nov 22 08:17:01 crc kubenswrapper[4735]: W1122 08:17:01.840914 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d914059_b35d_4dd7_9e70_d9bf8877c0c0.slice/crio-bd16033fb3c84a5a96767a18844c1cef17d11683e28d55540feaa2a3d00b2d2d WatchSource:0}: Error finding container bd16033fb3c84a5a96767a18844c1cef17d11683e28d55540feaa2a3d00b2d2d: Status 404 returned error can't find the container with id bd16033fb3c84a5a96767a18844c1cef17d11683e28d55540feaa2a3d00b2d2d Nov 22 08:17:02 crc kubenswrapper[4735]: I1122 08:17:02.377947 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7e069fd-4e58-471d-bc39-4c759edd9c9f" (UID: "a7e069fd-4e58-471d-bc39-4c759edd9c9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:02 crc kubenswrapper[4735]: I1122 08:17:02.462172 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e069fd-4e58-471d-bc39-4c759edd9c9f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:02 crc kubenswrapper[4735]: I1122 08:17:02.484351 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t5vpq"] Nov 22 08:17:02 crc kubenswrapper[4735]: I1122 08:17:02.494553 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t5vpq"] Nov 22 08:17:02 crc kubenswrapper[4735]: I1122 08:17:02.540536 4735 generic.go:334] "Generic (PLEG): container finished" podID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerID="b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283" exitCode=0 Nov 22 08:17:02 crc kubenswrapper[4735]: I1122 08:17:02.540589 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pjm5" event={"ID":"8d914059-b35d-4dd7-9e70-d9bf8877c0c0","Type":"ContainerDied","Data":"b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283"} Nov 22 08:17:02 crc kubenswrapper[4735]: I1122 08:17:02.540615 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pjm5" event={"ID":"8d914059-b35d-4dd7-9e70-d9bf8877c0c0","Type":"ContainerStarted","Data":"bd16033fb3c84a5a96767a18844c1cef17d11683e28d55540feaa2a3d00b2d2d"} Nov 22 08:17:03 crc kubenswrapper[4735]: I1122 08:17:03.273436 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" path="/var/lib/kubelet/pods/a7e069fd-4e58-471d-bc39-4c759edd9c9f/volumes" Nov 22 08:17:03 crc kubenswrapper[4735]: I1122 08:17:03.554825 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pjm5" event={"ID":"8d914059-b35d-4dd7-9e70-d9bf8877c0c0","Type":"ContainerStarted","Data":"a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f"} Nov 22 08:17:04 crc kubenswrapper[4735]: I1122 08:17:04.566206 4735 generic.go:334] "Generic (PLEG): container finished" podID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerID="a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f" exitCode=0 Nov 22 08:17:04 crc kubenswrapper[4735]: I1122 08:17:04.566298 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pjm5" event={"ID":"8d914059-b35d-4dd7-9e70-d9bf8877c0c0","Type":"ContainerDied","Data":"a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f"} Nov 22 08:17:05 crc kubenswrapper[4735]: I1122 08:17:05.579535 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pjm5" event={"ID":"8d914059-b35d-4dd7-9e70-d9bf8877c0c0","Type":"ContainerStarted","Data":"a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0"} Nov 22 08:17:05 crc kubenswrapper[4735]: I1122 08:17:05.615163 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6pjm5" podStartSLOduration=2.176673955 podStartE2EDuration="4.615142475s" podCreationTimestamp="2025-11-22 08:17:01 +0000 UTC" firstStartedPulling="2025-11-22 08:17:02.54382394 +0000 UTC m=+844.148162535" lastFinishedPulling="2025-11-22 08:17:04.98229245 +0000 UTC m=+846.586631055" observedRunningTime="2025-11-22 08:17:05.614240271 +0000 UTC m=+847.218578916" watchObservedRunningTime="2025-11-22 08:17:05.615142475 +0000 UTC m=+847.219481080" Nov 22 08:17:11 crc kubenswrapper[4735]: I1122 08:17:11.385696 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:11 crc kubenswrapper[4735]: I1122 08:17:11.386223 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:11 crc kubenswrapper[4735]: I1122 08:17:11.440697 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:11 crc kubenswrapper[4735]: I1122 08:17:11.654666 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:11 crc kubenswrapper[4735]: I1122 08:17:11.703295 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6pjm5"] Nov 22 08:17:13 crc kubenswrapper[4735]: I1122 08:17:13.632657 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6pjm5" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerName="registry-server" containerID="cri-o://a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0" gracePeriod=2 Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.613485 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.647290 4735 generic.go:334] "Generic (PLEG): container finished" podID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerID="a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0" exitCode=0 Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.647346 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6pjm5" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.647348 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pjm5" event={"ID":"8d914059-b35d-4dd7-9e70-d9bf8877c0c0","Type":"ContainerDied","Data":"a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0"} Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.647400 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pjm5" event={"ID":"8d914059-b35d-4dd7-9e70-d9bf8877c0c0","Type":"ContainerDied","Data":"bd16033fb3c84a5a96767a18844c1cef17d11683e28d55540feaa2a3d00b2d2d"} Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.647423 4735 scope.go:117] "RemoveContainer" containerID="a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.687642 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt2q7\" (UniqueName: \"kubernetes.io/projected/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-kube-api-access-qt2q7\") pod \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.687756 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-catalog-content\") pod \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.687845 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-utilities\") pod \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\" (UID: \"8d914059-b35d-4dd7-9e70-d9bf8877c0c0\") " Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.689104 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-utilities" (OuterVolumeSpecName: "utilities") pod "8d914059-b35d-4dd7-9e70-d9bf8877c0c0" (UID: "8d914059-b35d-4dd7-9e70-d9bf8877c0c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.702310 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-kube-api-access-qt2q7" (OuterVolumeSpecName: "kube-api-access-qt2q7") pod "8d914059-b35d-4dd7-9e70-d9bf8877c0c0" (UID: "8d914059-b35d-4dd7-9e70-d9bf8877c0c0"). InnerVolumeSpecName "kube-api-access-qt2q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.714777 4735 scope.go:117] "RemoveContainer" containerID="a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.751942 4735 scope.go:117] "RemoveContainer" containerID="b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.775658 4735 scope.go:117] "RemoveContainer" containerID="a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0" Nov 22 08:17:14 crc kubenswrapper[4735]: E1122 08:17:14.776494 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0\": container with ID starting with a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0 not found: ID does not exist" containerID="a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.776603 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0"} err="failed to get container status \"a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0\": rpc error: code = NotFound desc = could not find container \"a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0\": container with ID starting with a3dbf4e9526c61f68e6952ed2517d2fdd3836cf102ecd87c2691683127fe43c0 not found: ID does not exist" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.776643 4735 scope.go:117] "RemoveContainer" containerID="a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f" Nov 22 08:17:14 crc kubenswrapper[4735]: E1122 08:17:14.777064 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f\": container with ID starting with a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f not found: ID does not exist" containerID="a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.777107 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f"} err="failed to get container status \"a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f\": rpc error: code = NotFound desc = could not find container \"a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f\": container with ID starting with a52bd7dd3d2e519723f90d56099995a766013eb1c14895c5da1eef7d95d1261f not found: ID does not exist" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.777131 4735 scope.go:117] "RemoveContainer" containerID="b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283" Nov 22 08:17:14 crc kubenswrapper[4735]: E1122 08:17:14.777447 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283\": container with ID starting with b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283 not found: ID does not exist" containerID="b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.777505 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283"} err="failed to get container status \"b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283\": rpc error: code = NotFound desc = could not find container \"b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283\": container with ID starting with b14acfaf814087ec5d7e6f174af8d150f610cca3162e914f2fe089062e42e283 not found: ID does not exist" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.789985 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.790027 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt2q7\" (UniqueName: \"kubernetes.io/projected/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-kube-api-access-qt2q7\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.799673 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d914059-b35d-4dd7-9e70-d9bf8877c0c0" (UID: "8d914059-b35d-4dd7-9e70-d9bf8877c0c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.891310 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d914059-b35d-4dd7-9e70-d9bf8877c0c0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:14 crc kubenswrapper[4735]: I1122 08:17:14.990817 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6pjm5"] Nov 22 08:17:15 crc kubenswrapper[4735]: I1122 08:17:15.000309 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6pjm5"] Nov 22 08:17:15 crc kubenswrapper[4735]: I1122 08:17:15.272409 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" path="/var/lib/kubelet/pods/8d914059-b35d-4dd7-9e70-d9bf8877c0c0/volumes" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.990541 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9"] Nov 22 08:17:20 crc kubenswrapper[4735]: E1122 08:17:20.991104 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerName="extract-utilities" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.991120 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerName="extract-utilities" Nov 22 08:17:20 crc kubenswrapper[4735]: E1122 08:17:20.991132 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerName="extract-utilities" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.991139 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerName="extract-utilities" Nov 22 08:17:20 crc kubenswrapper[4735]: E1122 08:17:20.991154 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerName="extract-content" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.991163 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerName="extract-content" Nov 22 08:17:20 crc kubenswrapper[4735]: E1122 08:17:20.991179 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerName="registry-server" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.991185 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerName="registry-server" Nov 22 08:17:20 crc kubenswrapper[4735]: E1122 08:17:20.991200 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerName="registry-server" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.991207 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerName="registry-server" Nov 22 08:17:20 crc kubenswrapper[4735]: E1122 08:17:20.991223 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerName="extract-content" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.991229 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerName="extract-content" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.991392 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d914059-b35d-4dd7-9e70-d9bf8877c0c0" containerName="registry-server" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.991411 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e069fd-4e58-471d-bc39-4c759edd9c9f" containerName="registry-server" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.992445 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:20 crc kubenswrapper[4735]: I1122 08:17:20.994894 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.010875 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9"] Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.092660 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfd48\" (UniqueName: \"kubernetes.io/projected/bd3ee390-fac8-4865-9be5-469cc9c2d791-kube-api-access-kfd48\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.092733 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.092756 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.194423 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.194500 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.194607 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfd48\" (UniqueName: \"kubernetes.io/projected/bd3ee390-fac8-4865-9be5-469cc9c2d791-kube-api-access-kfd48\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.194985 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.195235 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.219682 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfd48\" (UniqueName: \"kubernetes.io/projected/bd3ee390-fac8-4865-9be5-469cc9c2d791-kube-api-access-kfd48\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.311699 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:21 crc kubenswrapper[4735]: I1122 08:17:21.781981 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9"] Nov 22 08:17:22 crc kubenswrapper[4735]: I1122 08:17:22.715118 4735 generic.go:334] "Generic (PLEG): container finished" podID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerID="698c43dc12c9814251f370ca91a37958ccbaa242ed62702e1c2c4e9bd7c2a8de" exitCode=0 Nov 22 08:17:22 crc kubenswrapper[4735]: I1122 08:17:22.716546 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" event={"ID":"bd3ee390-fac8-4865-9be5-469cc9c2d791","Type":"ContainerDied","Data":"698c43dc12c9814251f370ca91a37958ccbaa242ed62702e1c2c4e9bd7c2a8de"} Nov 22 08:17:22 crc kubenswrapper[4735]: I1122 08:17:22.718744 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" event={"ID":"bd3ee390-fac8-4865-9be5-469cc9c2d791","Type":"ContainerStarted","Data":"100300ef716edad6fdd8847ac8a541587ae706e81cb42a03f46a3d061540295e"} Nov 22 08:17:24 crc kubenswrapper[4735]: I1122 08:17:24.736942 4735 generic.go:334] "Generic (PLEG): container finished" podID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerID="71d9ca733a3dfcef618cd39d3c5daf429962d6b1a8c0014fecf6ed59a9316c05" exitCode=0 Nov 22 08:17:24 crc kubenswrapper[4735]: I1122 08:17:24.736989 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" event={"ID":"bd3ee390-fac8-4865-9be5-469cc9c2d791","Type":"ContainerDied","Data":"71d9ca733a3dfcef618cd39d3c5daf429962d6b1a8c0014fecf6ed59a9316c05"} Nov 22 08:17:25 crc kubenswrapper[4735]: I1122 08:17:25.748146 4735 generic.go:334] "Generic (PLEG): container finished" podID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerID="6e17e26ab052e82e4ba6a274e3c86b2a77fc090efdb7e4a116b4c5270f31ea21" exitCode=0 Nov 22 08:17:25 crc kubenswrapper[4735]: I1122 08:17:25.748210 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" event={"ID":"bd3ee390-fac8-4865-9be5-469cc9c2d791","Type":"ContainerDied","Data":"6e17e26ab052e82e4ba6a274e3c86b2a77fc090efdb7e4a116b4c5270f31ea21"} Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.088605 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.211938 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-util\") pod \"bd3ee390-fac8-4865-9be5-469cc9c2d791\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.212094 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfd48\" (UniqueName: \"kubernetes.io/projected/bd3ee390-fac8-4865-9be5-469cc9c2d791-kube-api-access-kfd48\") pod \"bd3ee390-fac8-4865-9be5-469cc9c2d791\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.212151 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-bundle\") pod \"bd3ee390-fac8-4865-9be5-469cc9c2d791\" (UID: \"bd3ee390-fac8-4865-9be5-469cc9c2d791\") " Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.212898 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-bundle" (OuterVolumeSpecName: "bundle") pod "bd3ee390-fac8-4865-9be5-469cc9c2d791" (UID: "bd3ee390-fac8-4865-9be5-469cc9c2d791"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.218053 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd3ee390-fac8-4865-9be5-469cc9c2d791-kube-api-access-kfd48" (OuterVolumeSpecName: "kube-api-access-kfd48") pod "bd3ee390-fac8-4865-9be5-469cc9c2d791" (UID: "bd3ee390-fac8-4865-9be5-469cc9c2d791"). InnerVolumeSpecName "kube-api-access-kfd48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.227605 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-util" (OuterVolumeSpecName: "util") pod "bd3ee390-fac8-4865-9be5-469cc9c2d791" (UID: "bd3ee390-fac8-4865-9be5-469cc9c2d791"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.314579 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.314813 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfd48\" (UniqueName: \"kubernetes.io/projected/bd3ee390-fac8-4865-9be5-469cc9c2d791-kube-api-access-kfd48\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.314877 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd3ee390-fac8-4865-9be5-469cc9c2d791-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.765315 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" event={"ID":"bd3ee390-fac8-4865-9be5-469cc9c2d791","Type":"ContainerDied","Data":"100300ef716edad6fdd8847ac8a541587ae706e81cb42a03f46a3d061540295e"} Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.765356 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="100300ef716edad6fdd8847ac8a541587ae706e81cb42a03f46a3d061540295e" Nov 22 08:17:27 crc kubenswrapper[4735]: I1122 08:17:27.765370 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.444363 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-9gq6d"] Nov 22 08:17:30 crc kubenswrapper[4735]: E1122 08:17:30.445151 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerName="extract" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.445165 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerName="extract" Nov 22 08:17:30 crc kubenswrapper[4735]: E1122 08:17:30.445183 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerName="pull" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.445190 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerName="pull" Nov 22 08:17:30 crc kubenswrapper[4735]: E1122 08:17:30.445219 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerName="util" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.445227 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerName="util" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.445384 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd3ee390-fac8-4865-9be5-469cc9c2d791" containerName="extract" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.446031 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-9gq6d" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.449023 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-kfklp" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.449386 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.453835 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.469969 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-9gq6d"] Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.581167 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh7bj\" (UniqueName: \"kubernetes.io/projected/36088127-63b1-40be-8456-294d89283c41-kube-api-access-xh7bj\") pod \"nmstate-operator-557fdffb88-9gq6d\" (UID: \"36088127-63b1-40be-8456-294d89283c41\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-9gq6d" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.685905 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh7bj\" (UniqueName: \"kubernetes.io/projected/36088127-63b1-40be-8456-294d89283c41-kube-api-access-xh7bj\") pod \"nmstate-operator-557fdffb88-9gq6d\" (UID: \"36088127-63b1-40be-8456-294d89283c41\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-9gq6d" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.709107 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh7bj\" (UniqueName: \"kubernetes.io/projected/36088127-63b1-40be-8456-294d89283c41-kube-api-access-xh7bj\") pod \"nmstate-operator-557fdffb88-9gq6d\" (UID: \"36088127-63b1-40be-8456-294d89283c41\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-9gq6d" Nov 22 08:17:30 crc kubenswrapper[4735]: I1122 08:17:30.764978 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-9gq6d" Nov 22 08:17:31 crc kubenswrapper[4735]: I1122 08:17:31.254981 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-9gq6d"] Nov 22 08:17:31 crc kubenswrapper[4735]: I1122 08:17:31.795894 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-9gq6d" event={"ID":"36088127-63b1-40be-8456-294d89283c41","Type":"ContainerStarted","Data":"2a422fd36db9dcc0c6b7efde2f6f7e7f3a18ddb314105eddfa8a17660019c56a"} Nov 22 08:17:33 crc kubenswrapper[4735]: I1122 08:17:33.811639 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-9gq6d" event={"ID":"36088127-63b1-40be-8456-294d89283c41","Type":"ContainerStarted","Data":"a9324d28a5b0381c9b2dc7342832b451f9d1dafe3736610fbe643a74b14844f6"} Nov 22 08:17:33 crc kubenswrapper[4735]: I1122 08:17:33.827508 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-9gq6d" podStartSLOduration=1.57992467 podStartE2EDuration="3.827488865s" podCreationTimestamp="2025-11-22 08:17:30 +0000 UTC" firstStartedPulling="2025-11-22 08:17:31.261811237 +0000 UTC m=+872.866149842" lastFinishedPulling="2025-11-22 08:17:33.509375392 +0000 UTC m=+875.113714037" observedRunningTime="2025-11-22 08:17:33.827121985 +0000 UTC m=+875.431460600" watchObservedRunningTime="2025-11-22 08:17:33.827488865 +0000 UTC m=+875.431827470" Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.855206 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh"] Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.856739 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.860889 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-mx84v" Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.866310 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-swwps"] Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.868592 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.870590 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.872566 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh"] Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.882512 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-swwps"] Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.917528 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-jslbs"] Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.918433 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.965879 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpcgz\" (UniqueName: \"kubernetes.io/projected/c2e387de-32b1-4259-9301-0181b08ab88d-kube-api-access-gpcgz\") pod \"nmstate-metrics-5dcf9c57c5-hnkkh\" (UID: \"c2e387de-32b1-4259-9301-0181b08ab88d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.965925 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw4cx\" (UniqueName: \"kubernetes.io/projected/fd488627-bc81-4723-b465-b39ff151d0d6-kube-api-access-hw4cx\") pod \"nmstate-webhook-6b89b748d8-swwps\" (UID: \"fd488627-bc81-4723-b465-b39ff151d0d6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:34 crc kubenswrapper[4735]: I1122 08:17:34.965981 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fd488627-bc81-4723-b465-b39ff151d0d6-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-swwps\" (UID: \"fd488627-bc81-4723-b465-b39ff151d0d6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.042522 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7"] Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.043640 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.045951 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.046117 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-fkk9n" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.049407 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.051279 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7"] Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.067549 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-dbus-socket\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.067612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fd488627-bc81-4723-b465-b39ff151d0d6-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-swwps\" (UID: \"fd488627-bc81-4723-b465-b39ff151d0d6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.067639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-nmstate-lock\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.067692 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxfcx\" (UniqueName: \"kubernetes.io/projected/45633081-910f-45c1-a095-21db07a7cc78-kube-api-access-sxfcx\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.067714 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-ovs-socket\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.067753 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpcgz\" (UniqueName: \"kubernetes.io/projected/c2e387de-32b1-4259-9301-0181b08ab88d-kube-api-access-gpcgz\") pod \"nmstate-metrics-5dcf9c57c5-hnkkh\" (UID: \"c2e387de-32b1-4259-9301-0181b08ab88d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.067775 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw4cx\" (UniqueName: \"kubernetes.io/projected/fd488627-bc81-4723-b465-b39ff151d0d6-kube-api-access-hw4cx\") pod \"nmstate-webhook-6b89b748d8-swwps\" (UID: \"fd488627-bc81-4723-b465-b39ff151d0d6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.073728 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fd488627-bc81-4723-b465-b39ff151d0d6-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-swwps\" (UID: \"fd488627-bc81-4723-b465-b39ff151d0d6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.098052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpcgz\" (UniqueName: \"kubernetes.io/projected/c2e387de-32b1-4259-9301-0181b08ab88d-kube-api-access-gpcgz\") pod \"nmstate-metrics-5dcf9c57c5-hnkkh\" (UID: \"c2e387de-32b1-4259-9301-0181b08ab88d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.100298 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw4cx\" (UniqueName: \"kubernetes.io/projected/fd488627-bc81-4723-b465-b39ff151d0d6-kube-api-access-hw4cx\") pod \"nmstate-webhook-6b89b748d8-swwps\" (UID: \"fd488627-bc81-4723-b465-b39ff151d0d6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.169767 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxfcx\" (UniqueName: \"kubernetes.io/projected/45633081-910f-45c1-a095-21db07a7cc78-kube-api-access-sxfcx\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.169832 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-ovs-socket\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.169873 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhbff\" (UniqueName: \"kubernetes.io/projected/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-kube-api-access-hhbff\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.169907 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.169967 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.170020 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-dbus-socket\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.170074 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-nmstate-lock\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.170179 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-nmstate-lock\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.170225 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-ovs-socket\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.170449 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/45633081-910f-45c1-a095-21db07a7cc78-dbus-socket\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.173163 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.196140 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxfcx\" (UniqueName: \"kubernetes.io/projected/45633081-910f-45c1-a095-21db07a7cc78-kube-api-access-sxfcx\") pod \"nmstate-handler-jslbs\" (UID: \"45633081-910f-45c1-a095-21db07a7cc78\") " pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.205126 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.238012 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.272872 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhbff\" (UniqueName: \"kubernetes.io/projected/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-kube-api-access-hhbff\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.272928 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.272996 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.274073 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.278064 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.294325 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6d7db65f59-c579x"] Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.295317 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d7db65f59-c579x"] Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.295415 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.311252 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhbff\" (UniqueName: \"kubernetes.io/projected/3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd-kube-api-access-hhbff\") pod \"nmstate-console-plugin-5874bd7bc5-4pgd7\" (UID: \"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.377235 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-console-config\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.377283 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-oauth-serving-cert\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.377323 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-service-ca\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.377371 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqtcp\" (UniqueName: \"kubernetes.io/projected/847def28-51ed-4e0d-8629-79c707aebb60-kube-api-access-fqtcp\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.377430 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-serving-cert\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.377496 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-trusted-ca-bundle\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.377519 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-oauth-config\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.466664 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.478432 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqtcp\" (UniqueName: \"kubernetes.io/projected/847def28-51ed-4e0d-8629-79c707aebb60-kube-api-access-fqtcp\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.478532 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-serving-cert\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.478558 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-trusted-ca-bundle\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.478580 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-oauth-config\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.478612 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-console-config\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.478637 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-oauth-serving-cert\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.478671 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-service-ca\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.479503 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-service-ca\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.483286 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-console-config\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.483624 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-trusted-ca-bundle\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.483941 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-oauth-serving-cert\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.490347 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-serving-cert\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.490921 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-oauth-config\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.500393 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqtcp\" (UniqueName: \"kubernetes.io/projected/847def28-51ed-4e0d-8629-79c707aebb60-kube-api-access-fqtcp\") pod \"console-6d7db65f59-c579x\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.631130 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.714131 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh"] Nov 22 08:17:35 crc kubenswrapper[4735]: W1122 08:17:35.728445 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2e387de_32b1_4259_9301_0181b08ab88d.slice/crio-a29888e1e7125f0b946085956fa68e66e2fad3c79a1097c41807a68d36f84522 WatchSource:0}: Error finding container a29888e1e7125f0b946085956fa68e66e2fad3c79a1097c41807a68d36f84522: Status 404 returned error can't find the container with id a29888e1e7125f0b946085956fa68e66e2fad3c79a1097c41807a68d36f84522 Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.786905 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-swwps"] Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.825830 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jslbs" event={"ID":"45633081-910f-45c1-a095-21db07a7cc78","Type":"ContainerStarted","Data":"4dbb92ae1f7303045be404cb04abb77c8f105c90f47330b9916d8847acf2534f"} Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.827694 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" event={"ID":"c2e387de-32b1-4259-9301-0181b08ab88d","Type":"ContainerStarted","Data":"a29888e1e7125f0b946085956fa68e66e2fad3c79a1097c41807a68d36f84522"} Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.828531 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" event={"ID":"fd488627-bc81-4723-b465-b39ff151d0d6","Type":"ContainerStarted","Data":"0d9aca503c0f58a2c794d3aa7772fe43da151ff7849c5f7aa0ead89db98b461f"} Nov 22 08:17:35 crc kubenswrapper[4735]: I1122 08:17:35.920853 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7"] Nov 22 08:17:35 crc kubenswrapper[4735]: W1122 08:17:35.925059 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a0c81e9_d65f_4b9e_8ea9_b1dc9793cabd.slice/crio-8f274ac59d39df6bef4c440fb594a3434ae5ee576230cb6da2778c280a26ae38 WatchSource:0}: Error finding container 8f274ac59d39df6bef4c440fb594a3434ae5ee576230cb6da2778c280a26ae38: Status 404 returned error can't find the container with id 8f274ac59d39df6bef4c440fb594a3434ae5ee576230cb6da2778c280a26ae38 Nov 22 08:17:36 crc kubenswrapper[4735]: I1122 08:17:36.063937 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d7db65f59-c579x"] Nov 22 08:17:36 crc kubenswrapper[4735]: I1122 08:17:36.837036 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d7db65f59-c579x" event={"ID":"847def28-51ed-4e0d-8629-79c707aebb60","Type":"ContainerStarted","Data":"7c333a1cc524ddd1fe128428db18e84333cdb9a5bf8934630e4854f6454a3c0c"} Nov 22 08:17:36 crc kubenswrapper[4735]: I1122 08:17:36.837533 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d7db65f59-c579x" event={"ID":"847def28-51ed-4e0d-8629-79c707aebb60","Type":"ContainerStarted","Data":"bb7cd2942fc23837733d386015a0bbadf60a31edacfa3c0c9ea151810bcc79ed"} Nov 22 08:17:36 crc kubenswrapper[4735]: I1122 08:17:36.839750 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" event={"ID":"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd","Type":"ContainerStarted","Data":"8f274ac59d39df6bef4c440fb594a3434ae5ee576230cb6da2778c280a26ae38"} Nov 22 08:17:36 crc kubenswrapper[4735]: I1122 08:17:36.862410 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6d7db65f59-c579x" podStartSLOduration=1.8623919930000001 podStartE2EDuration="1.862391993s" podCreationTimestamp="2025-11-22 08:17:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:17:36.857072827 +0000 UTC m=+878.461411452" watchObservedRunningTime="2025-11-22 08:17:36.862391993 +0000 UTC m=+878.466730588" Nov 22 08:17:40 crc kubenswrapper[4735]: I1122 08:17:40.869601 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jslbs" event={"ID":"45633081-910f-45c1-a095-21db07a7cc78","Type":"ContainerStarted","Data":"0e505f62790074b2e4b743295e78b2844cc26ae3263b236ecb6658a875c19d97"} Nov 22 08:17:40 crc kubenswrapper[4735]: I1122 08:17:40.871651 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" event={"ID":"c2e387de-32b1-4259-9301-0181b08ab88d","Type":"ContainerStarted","Data":"09a1677f2b24f6ad09318881452dc629e15ad887cbcfee190401ffc35fd430a4"} Nov 22 08:17:40 crc kubenswrapper[4735]: I1122 08:17:40.879161 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" event={"ID":"3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd","Type":"ContainerStarted","Data":"95b0317ecc0c197b11e9e2bcb2d60e5a749382ae17fb688b61c2a7829e341960"} Nov 22 08:17:40 crc kubenswrapper[4735]: I1122 08:17:40.881498 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" event={"ID":"fd488627-bc81-4723-b465-b39ff151d0d6","Type":"ContainerStarted","Data":"072f6603362b14ff7d2b3158a233c4e2a075d9f92570ee22d6fc09b81f0528f6"} Nov 22 08:17:40 crc kubenswrapper[4735]: I1122 08:17:40.882408 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:17:40 crc kubenswrapper[4735]: I1122 08:17:40.901134 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-jslbs" podStartSLOduration=2.579117599 podStartE2EDuration="6.901115361s" podCreationTimestamp="2025-11-22 08:17:34 +0000 UTC" firstStartedPulling="2025-11-22 08:17:35.295438097 +0000 UTC m=+876.899776702" lastFinishedPulling="2025-11-22 08:17:39.617435859 +0000 UTC m=+881.221774464" observedRunningTime="2025-11-22 08:17:40.894293735 +0000 UTC m=+882.498632360" watchObservedRunningTime="2025-11-22 08:17:40.901115361 +0000 UTC m=+882.505453976" Nov 22 08:17:40 crc kubenswrapper[4735]: I1122 08:17:40.923100 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" podStartSLOduration=3.100378308 podStartE2EDuration="6.923083701s" podCreationTimestamp="2025-11-22 08:17:34 +0000 UTC" firstStartedPulling="2025-11-22 08:17:35.794830009 +0000 UTC m=+877.399168614" lastFinishedPulling="2025-11-22 08:17:39.617535402 +0000 UTC m=+881.221874007" observedRunningTime="2025-11-22 08:17:40.920559402 +0000 UTC m=+882.524898027" watchObservedRunningTime="2025-11-22 08:17:40.923083701 +0000 UTC m=+882.527422306" Nov 22 08:17:40 crc kubenswrapper[4735]: I1122 08:17:40.958209 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-4pgd7" podStartSLOduration=1.692583967 podStartE2EDuration="5.958182909s" podCreationTimestamp="2025-11-22 08:17:35 +0000 UTC" firstStartedPulling="2025-11-22 08:17:35.92780696 +0000 UTC m=+877.532145565" lastFinishedPulling="2025-11-22 08:17:40.193405902 +0000 UTC m=+881.797744507" observedRunningTime="2025-11-22 08:17:40.957729486 +0000 UTC m=+882.562068121" watchObservedRunningTime="2025-11-22 08:17:40.958182909 +0000 UTC m=+882.562521544" Nov 22 08:17:41 crc kubenswrapper[4735]: I1122 08:17:41.889218 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:44 crc kubenswrapper[4735]: I1122 08:17:44.934560 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" event={"ID":"c2e387de-32b1-4259-9301-0181b08ab88d","Type":"ContainerStarted","Data":"74761aa01eaedc41abd7385a1a4a3f403b9c7f92b4f509adca1e99bca7a7b1b8"} Nov 22 08:17:44 crc kubenswrapper[4735]: I1122 08:17:44.958772 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hnkkh" podStartSLOduration=2.505952643 podStartE2EDuration="10.958753667s" podCreationTimestamp="2025-11-22 08:17:34 +0000 UTC" firstStartedPulling="2025-11-22 08:17:35.731980884 +0000 UTC m=+877.336319489" lastFinishedPulling="2025-11-22 08:17:44.184781908 +0000 UTC m=+885.789120513" observedRunningTime="2025-11-22 08:17:44.95376273 +0000 UTC m=+886.558101335" watchObservedRunningTime="2025-11-22 08:17:44.958753667 +0000 UTC m=+886.563092272" Nov 22 08:17:45 crc kubenswrapper[4735]: I1122 08:17:45.271855 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-jslbs" Nov 22 08:17:45 crc kubenswrapper[4735]: I1122 08:17:45.631875 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:45 crc kubenswrapper[4735]: I1122 08:17:45.631983 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:45 crc kubenswrapper[4735]: I1122 08:17:45.637001 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:45 crc kubenswrapper[4735]: I1122 08:17:45.946204 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:17:46 crc kubenswrapper[4735]: I1122 08:17:46.016800 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-56b67fb7dd-w5qgb"] Nov 22 08:17:46 crc kubenswrapper[4735]: I1122 08:17:46.131233 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:17:46 crc kubenswrapper[4735]: I1122 08:17:46.131332 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:17:55 crc kubenswrapper[4735]: I1122 08:17:55.213070 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" Nov 22 08:18:11 crc kubenswrapper[4735]: I1122 08:18:11.069596 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-56b67fb7dd-w5qgb" podUID="4c2b6c8e-3c56-4e20-a28b-182129e41afd" containerName="console" containerID="cri-o://c86af4eaaadec71e246e5cda1aa5bca586583afdd74b0fee27518b1aa9384c33" gracePeriod=15 Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.131632 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-56b67fb7dd-w5qgb_4c2b6c8e-3c56-4e20-a28b-182129e41afd/console/0.log" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.131958 4735 generic.go:334] "Generic (PLEG): container finished" podID="4c2b6c8e-3c56-4e20-a28b-182129e41afd" containerID="c86af4eaaadec71e246e5cda1aa5bca586583afdd74b0fee27518b1aa9384c33" exitCode=2 Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.131993 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-56b67fb7dd-w5qgb" event={"ID":"4c2b6c8e-3c56-4e20-a28b-182129e41afd","Type":"ContainerDied","Data":"c86af4eaaadec71e246e5cda1aa5bca586583afdd74b0fee27518b1aa9384c33"} Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.132032 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-56b67fb7dd-w5qgb" event={"ID":"4c2b6c8e-3c56-4e20-a28b-182129e41afd","Type":"ContainerDied","Data":"cedcbd5dd91e4f2a0afa57e85f953ee407428f4cce3f5536ca5fcd594ca69b3f"} Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.132047 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cedcbd5dd91e4f2a0afa57e85f953ee407428f4cce3f5536ca5fcd594ca69b3f" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.149083 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-56b67fb7dd-w5qgb_4c2b6c8e-3c56-4e20-a28b-182129e41afd/console/0.log" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.149163 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.302537 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-service-ca\") pod \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.302976 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-trusted-ca-bundle\") pod \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.303473 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-service-ca" (OuterVolumeSpecName: "service-ca") pod "4c2b6c8e-3c56-4e20-a28b-182129e41afd" (UID: "4c2b6c8e-3c56-4e20-a28b-182129e41afd"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.303501 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4c2b6c8e-3c56-4e20-a28b-182129e41afd" (UID: "4c2b6c8e-3c56-4e20-a28b-182129e41afd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.303622 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-config\") pod \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.303691 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-oauth-config\") pod \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.303771 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-oauth-serving-cert\") pod \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.303808 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-serving-cert\") pod \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.303832 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz9sq\" (UniqueName: \"kubernetes.io/projected/4c2b6c8e-3c56-4e20-a28b-182129e41afd-kube-api-access-kz9sq\") pod \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\" (UID: \"4c2b6c8e-3c56-4e20-a28b-182129e41afd\") " Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.304313 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-config" (OuterVolumeSpecName: "console-config") pod "4c2b6c8e-3c56-4e20-a28b-182129e41afd" (UID: "4c2b6c8e-3c56-4e20-a28b-182129e41afd"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.304512 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.304527 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.304538 4735 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.305596 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4c2b6c8e-3c56-4e20-a28b-182129e41afd" (UID: "4c2b6c8e-3c56-4e20-a28b-182129e41afd"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.310516 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2b6c8e-3c56-4e20-a28b-182129e41afd-kube-api-access-kz9sq" (OuterVolumeSpecName: "kube-api-access-kz9sq") pod "4c2b6c8e-3c56-4e20-a28b-182129e41afd" (UID: "4c2b6c8e-3c56-4e20-a28b-182129e41afd"). InnerVolumeSpecName "kube-api-access-kz9sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.310882 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4c2b6c8e-3c56-4e20-a28b-182129e41afd" (UID: "4c2b6c8e-3c56-4e20-a28b-182129e41afd"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.313584 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4c2b6c8e-3c56-4e20-a28b-182129e41afd" (UID: "4c2b6c8e-3c56-4e20-a28b-182129e41afd"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.405386 4735 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.405419 4735 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c2b6c8e-3c56-4e20-a28b-182129e41afd-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.405428 4735 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2b6c8e-3c56-4e20-a28b-182129e41afd-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:12 crc kubenswrapper[4735]: I1122 08:18:12.405436 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz9sq\" (UniqueName: \"kubernetes.io/projected/4c2b6c8e-3c56-4e20-a28b-182129e41afd-kube-api-access-kz9sq\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:13 crc kubenswrapper[4735]: I1122 08:18:13.138484 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-56b67fb7dd-w5qgb" Nov 22 08:18:13 crc kubenswrapper[4735]: I1122 08:18:13.184071 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-56b67fb7dd-w5qgb"] Nov 22 08:18:13 crc kubenswrapper[4735]: I1122 08:18:13.190249 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-56b67fb7dd-w5qgb"] Nov 22 08:18:13 crc kubenswrapper[4735]: I1122 08:18:13.272431 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c2b6c8e-3c56-4e20-a28b-182129e41afd" path="/var/lib/kubelet/pods/4c2b6c8e-3c56-4e20-a28b-182129e41afd/volumes" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.020481 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59"] Nov 22 08:18:14 crc kubenswrapper[4735]: E1122 08:18:14.021080 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2b6c8e-3c56-4e20-a28b-182129e41afd" containerName="console" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.021105 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2b6c8e-3c56-4e20-a28b-182129e41afd" containerName="console" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.021257 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2b6c8e-3c56-4e20-a28b-182129e41afd" containerName="console" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.022352 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.024220 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.032053 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59"] Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.130592 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.130677 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gss9\" (UniqueName: \"kubernetes.io/projected/bb5e21db-7091-4338-8ea2-1a1f3d861984-kube-api-access-7gss9\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.130819 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.232495 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.232591 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.232631 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gss9\" (UniqueName: \"kubernetes.io/projected/bb5e21db-7091-4338-8ea2-1a1f3d861984-kube-api-access-7gss9\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.233016 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.233058 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.251311 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gss9\" (UniqueName: \"kubernetes.io/projected/bb5e21db-7091-4338-8ea2-1a1f3d861984-kube-api-access-7gss9\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.341076 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:14 crc kubenswrapper[4735]: I1122 08:18:14.821295 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59"] Nov 22 08:18:15 crc kubenswrapper[4735]: I1122 08:18:15.154977 4735 generic.go:334] "Generic (PLEG): container finished" podID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerID="4604b207a35059bbea1000086922823ee4382abfaf8d7c1547cef716d61e9b82" exitCode=0 Nov 22 08:18:15 crc kubenswrapper[4735]: I1122 08:18:15.155028 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" event={"ID":"bb5e21db-7091-4338-8ea2-1a1f3d861984","Type":"ContainerDied","Data":"4604b207a35059bbea1000086922823ee4382abfaf8d7c1547cef716d61e9b82"} Nov 22 08:18:15 crc kubenswrapper[4735]: I1122 08:18:15.155290 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" event={"ID":"bb5e21db-7091-4338-8ea2-1a1f3d861984","Type":"ContainerStarted","Data":"0fc838f0c3b69e6d99692e45a2c966b06efce793c22dbfe2f27fdb1203d97e1d"} Nov 22 08:18:15 crc kubenswrapper[4735]: I1122 08:18:15.156847 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:18:16 crc kubenswrapper[4735]: I1122 08:18:16.132109 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:18:16 crc kubenswrapper[4735]: I1122 08:18:16.132176 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:18:17 crc kubenswrapper[4735]: I1122 08:18:17.169491 4735 generic.go:334] "Generic (PLEG): container finished" podID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerID="5eac2b38e7eb34a838965d7f8701b91115e57b46fc060add4fe0a1a3a16795c2" exitCode=0 Nov 22 08:18:17 crc kubenswrapper[4735]: I1122 08:18:17.169543 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" event={"ID":"bb5e21db-7091-4338-8ea2-1a1f3d861984","Type":"ContainerDied","Data":"5eac2b38e7eb34a838965d7f8701b91115e57b46fc060add4fe0a1a3a16795c2"} Nov 22 08:18:18 crc kubenswrapper[4735]: I1122 08:18:18.180143 4735 generic.go:334] "Generic (PLEG): container finished" podID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerID="4b6f43157198782610cf5e648ab32e23f2190660d01c58c0b43864d178c6fbb8" exitCode=0 Nov 22 08:18:18 crc kubenswrapper[4735]: I1122 08:18:18.180198 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" event={"ID":"bb5e21db-7091-4338-8ea2-1a1f3d861984","Type":"ContainerDied","Data":"4b6f43157198782610cf5e648ab32e23f2190660d01c58c0b43864d178c6fbb8"} Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.462823 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.524614 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gss9\" (UniqueName: \"kubernetes.io/projected/bb5e21db-7091-4338-8ea2-1a1f3d861984-kube-api-access-7gss9\") pod \"bb5e21db-7091-4338-8ea2-1a1f3d861984\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.524898 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-bundle\") pod \"bb5e21db-7091-4338-8ea2-1a1f3d861984\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.524959 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-util\") pod \"bb5e21db-7091-4338-8ea2-1a1f3d861984\" (UID: \"bb5e21db-7091-4338-8ea2-1a1f3d861984\") " Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.525964 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-bundle" (OuterVolumeSpecName: "bundle") pod "bb5e21db-7091-4338-8ea2-1a1f3d861984" (UID: "bb5e21db-7091-4338-8ea2-1a1f3d861984"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.531827 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb5e21db-7091-4338-8ea2-1a1f3d861984-kube-api-access-7gss9" (OuterVolumeSpecName: "kube-api-access-7gss9") pod "bb5e21db-7091-4338-8ea2-1a1f3d861984" (UID: "bb5e21db-7091-4338-8ea2-1a1f3d861984"). InnerVolumeSpecName "kube-api-access-7gss9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.540224 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-util" (OuterVolumeSpecName: "util") pod "bb5e21db-7091-4338-8ea2-1a1f3d861984" (UID: "bb5e21db-7091-4338-8ea2-1a1f3d861984"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.627477 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.627828 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb5e21db-7091-4338-8ea2-1a1f3d861984-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:19 crc kubenswrapper[4735]: I1122 08:18:19.627841 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gss9\" (UniqueName: \"kubernetes.io/projected/bb5e21db-7091-4338-8ea2-1a1f3d861984-kube-api-access-7gss9\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:20 crc kubenswrapper[4735]: I1122 08:18:20.214167 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" event={"ID":"bb5e21db-7091-4338-8ea2-1a1f3d861984","Type":"ContainerDied","Data":"0fc838f0c3b69e6d99692e45a2c966b06efce793c22dbfe2f27fdb1203d97e1d"} Nov 22 08:18:20 crc kubenswrapper[4735]: I1122 08:18:20.214211 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fc838f0c3b69e6d99692e45a2c966b06efce793c22dbfe2f27fdb1203d97e1d" Nov 22 08:18:20 crc kubenswrapper[4735]: I1122 08:18:20.214265 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.620771 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5"] Nov 22 08:18:33 crc kubenswrapper[4735]: E1122 08:18:33.621364 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerName="util" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.621375 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerName="util" Nov 22 08:18:33 crc kubenswrapper[4735]: E1122 08:18:33.621389 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerName="pull" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.621395 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerName="pull" Nov 22 08:18:33 crc kubenswrapper[4735]: E1122 08:18:33.621406 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerName="extract" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.621413 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerName="extract" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.621541 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5e21db-7091-4338-8ea2-1a1f3d861984" containerName="extract" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.622019 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.626585 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.627693 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.634128 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.638043 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.638302 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-j6kr7" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.646195 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94a2a8b2-e02e-409a-9fa2-06016f577fb1-apiservice-cert\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.646692 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vhc4\" (UniqueName: \"kubernetes.io/projected/94a2a8b2-e02e-409a-9fa2-06016f577fb1-kube-api-access-9vhc4\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.646859 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94a2a8b2-e02e-409a-9fa2-06016f577fb1-webhook-cert\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.657272 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5"] Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.748710 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vhc4\" (UniqueName: \"kubernetes.io/projected/94a2a8b2-e02e-409a-9fa2-06016f577fb1-kube-api-access-9vhc4\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.748783 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94a2a8b2-e02e-409a-9fa2-06016f577fb1-webhook-cert\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.748829 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94a2a8b2-e02e-409a-9fa2-06016f577fb1-apiservice-cert\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.772026 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94a2a8b2-e02e-409a-9fa2-06016f577fb1-apiservice-cert\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.774133 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94a2a8b2-e02e-409a-9fa2-06016f577fb1-webhook-cert\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.778383 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vhc4\" (UniqueName: \"kubernetes.io/projected/94a2a8b2-e02e-409a-9fa2-06016f577fb1-kube-api-access-9vhc4\") pod \"metallb-operator-controller-manager-f4fbf457b-9k6b5\" (UID: \"94a2a8b2-e02e-409a-9fa2-06016f577fb1\") " pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:33 crc kubenswrapper[4735]: I1122 08:18:33.940960 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.035010 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-b67588696-x9fc4"] Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.036349 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.038692 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.038926 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-blf5k" Nov 22 08:18:34 crc kubenswrapper[4735]: W1122 08:18:34.039043 4735 reflector.go:561] object-"metallb-system"/"metallb-webhook-cert": failed to list *v1.Secret: secrets "metallb-webhook-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 22 08:18:34 crc kubenswrapper[4735]: E1122 08:18:34.039074 4735 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-webhook-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-webhook-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.050639 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-b67588696-x9fc4"] Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.052995 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61883f69-88c2-40a4-b62b-f258bbace22f-apiservice-cert\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.053084 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61883f69-88c2-40a4-b62b-f258bbace22f-webhook-cert\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.053127 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbzr4\" (UniqueName: \"kubernetes.io/projected/61883f69-88c2-40a4-b62b-f258bbace22f-kube-api-access-wbzr4\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.154342 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61883f69-88c2-40a4-b62b-f258bbace22f-webhook-cert\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.154401 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbzr4\" (UniqueName: \"kubernetes.io/projected/61883f69-88c2-40a4-b62b-f258bbace22f-kube-api-access-wbzr4\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.154474 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61883f69-88c2-40a4-b62b-f258bbace22f-apiservice-cert\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.161944 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61883f69-88c2-40a4-b62b-f258bbace22f-webhook-cert\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.172159 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61883f69-88c2-40a4-b62b-f258bbace22f-apiservice-cert\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.177604 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbzr4\" (UniqueName: \"kubernetes.io/projected/61883f69-88c2-40a4-b62b-f258bbace22f-kube-api-access-wbzr4\") pod \"metallb-operator-webhook-server-b67588696-x9fc4\" (UID: \"61883f69-88c2-40a4-b62b-f258bbace22f\") " pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.359975 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.478134 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5"] Nov 22 08:18:34 crc kubenswrapper[4735]: I1122 08:18:34.809566 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-b67588696-x9fc4"] Nov 22 08:18:34 crc kubenswrapper[4735]: W1122 08:18:34.817924 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61883f69_88c2_40a4_b62b_f258bbace22f.slice/crio-10ab3f68590f8542d68996234370d78eb78ed47066e7c950e54d53180150ccdf WatchSource:0}: Error finding container 10ab3f68590f8542d68996234370d78eb78ed47066e7c950e54d53180150ccdf: Status 404 returned error can't find the container with id 10ab3f68590f8542d68996234370d78eb78ed47066e7c950e54d53180150ccdf Nov 22 08:18:35 crc kubenswrapper[4735]: I1122 08:18:35.037116 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 22 08:18:35 crc kubenswrapper[4735]: I1122 08:18:35.337052 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" event={"ID":"61883f69-88c2-40a4-b62b-f258bbace22f","Type":"ContainerStarted","Data":"10ab3f68590f8542d68996234370d78eb78ed47066e7c950e54d53180150ccdf"} Nov 22 08:18:35 crc kubenswrapper[4735]: I1122 08:18:35.338921 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" event={"ID":"94a2a8b2-e02e-409a-9fa2-06016f577fb1","Type":"ContainerStarted","Data":"90b8e017d8126ffae64b673cefa2f354ef32b61cf29d8b6dee4246ea90bbf11d"} Nov 22 08:18:41 crc kubenswrapper[4735]: I1122 08:18:41.404975 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" event={"ID":"94a2a8b2-e02e-409a-9fa2-06016f577fb1","Type":"ContainerStarted","Data":"43baa028f9948e43fc302ee8cdeca3d72ba07279c173f9eb7dc2627857a154da"} Nov 22 08:18:41 crc kubenswrapper[4735]: I1122 08:18:41.405334 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:18:41 crc kubenswrapper[4735]: I1122 08:18:41.406831 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" event={"ID":"61883f69-88c2-40a4-b62b-f258bbace22f","Type":"ContainerStarted","Data":"0a13ba7198f9761c52ea58bbfc94e24dd321ff4b6b92a319df77694d83ec2ecf"} Nov 22 08:18:41 crc kubenswrapper[4735]: I1122 08:18:41.406952 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:41 crc kubenswrapper[4735]: I1122 08:18:41.430680 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" podStartSLOduration=2.783297654 podStartE2EDuration="8.430661842s" podCreationTimestamp="2025-11-22 08:18:33 +0000 UTC" firstStartedPulling="2025-11-22 08:18:34.489893565 +0000 UTC m=+936.094232170" lastFinishedPulling="2025-11-22 08:18:40.137257753 +0000 UTC m=+941.741596358" observedRunningTime="2025-11-22 08:18:41.424167995 +0000 UTC m=+943.028506600" watchObservedRunningTime="2025-11-22 08:18:41.430661842 +0000 UTC m=+943.035000447" Nov 22 08:18:41 crc kubenswrapper[4735]: I1122 08:18:41.453417 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" podStartSLOduration=2.093083194 podStartE2EDuration="7.453396618s" podCreationTimestamp="2025-11-22 08:18:34 +0000 UTC" firstStartedPulling="2025-11-22 08:18:34.821751414 +0000 UTC m=+936.426090019" lastFinishedPulling="2025-11-22 08:18:40.182064838 +0000 UTC m=+941.786403443" observedRunningTime="2025-11-22 08:18:41.448992235 +0000 UTC m=+943.053330840" watchObservedRunningTime="2025-11-22 08:18:41.453396618 +0000 UTC m=+943.057735223" Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.131982 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.132341 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.132392 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.133039 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ac3677e0ebb55329e2a912a3a81033db8ba05ef418621a9d574aafba4dff19d"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.133096 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://8ac3677e0ebb55329e2a912a3a81033db8ba05ef418621a9d574aafba4dff19d" gracePeriod=600 Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.444302 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="8ac3677e0ebb55329e2a912a3a81033db8ba05ef418621a9d574aafba4dff19d" exitCode=0 Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.444349 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"8ac3677e0ebb55329e2a912a3a81033db8ba05ef418621a9d574aafba4dff19d"} Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.444380 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"33c0e7ebc11dc73bdfa42a0d62c634dcd145b69c651d39f5221e590919439f72"} Nov 22 08:18:46 crc kubenswrapper[4735]: I1122 08:18:46.444403 4735 scope.go:117] "RemoveContainer" containerID="a3aab057b98bc5c57ec42d5f2111ef56b0a63d358e465d259bfc8b9b235ad36f" Nov 22 08:18:54 crc kubenswrapper[4735]: I1122 08:18:54.370171 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-b67588696-x9fc4" Nov 22 08:18:59 crc kubenswrapper[4735]: I1122 08:18:59.665720 4735 scope.go:117] "RemoveContainer" containerID="c86af4eaaadec71e246e5cda1aa5bca586583afdd74b0fee27518b1aa9384c33" Nov 22 08:19:13 crc kubenswrapper[4735]: I1122 08:19:13.945411 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.679515 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-bkh26"] Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.700373 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.710971 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.711284 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.711489 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-2qd2s" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.714580 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n"] Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.715685 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.717718 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.722686 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n"] Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.816420 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-klgsh"] Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819551 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-metrics-certs\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819609 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-sockets\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819648 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdkwc\" (UniqueName: \"kubernetes.io/projected/84fe49a3-baa7-45bf-aa7b-b92e55e61a1e-kube-api-access-kdkwc\") pod \"frr-k8s-webhook-server-6998585d5-wzm9n\" (UID: \"84fe49a3-baa7-45bf-aa7b-b92e55e61a1e\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819673 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-reloader\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819703 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/84fe49a3-baa7-45bf-aa7b-b92e55e61a1e-cert\") pod \"frr-k8s-webhook-server-6998585d5-wzm9n\" (UID: \"84fe49a3-baa7-45bf-aa7b-b92e55e61a1e\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819722 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljbmf\" (UniqueName: \"kubernetes.io/projected/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-kube-api-access-ljbmf\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819746 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-startup\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819803 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-metrics\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.819820 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-conf\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.820172 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-klgsh" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.824279 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.825788 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.825786 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.825789 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5jz6c" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.851867 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-2gwq7"] Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.852998 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.855083 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.864208 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-2gwq7"] Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921187 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-conf\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921243 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-metrics-certs\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921285 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-metrics-certs\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-sockets\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921354 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eddaaabc-450b-4c69-9e29-35f554b01893-metrics-certs\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921487 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eddaaabc-450b-4c69-9e29-35f554b01893-cert\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921595 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdkwc\" (UniqueName: \"kubernetes.io/projected/84fe49a3-baa7-45bf-aa7b-b92e55e61a1e-kube-api-access-kdkwc\") pod \"frr-k8s-webhook-server-6998585d5-wzm9n\" (UID: \"84fe49a3-baa7-45bf-aa7b-b92e55e61a1e\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921654 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-reloader\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921673 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-sockets\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921698 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e40c967d-3d49-4b06-8789-9b1abe501e80-metallb-excludel2\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921714 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-conf\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921724 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssf25\" (UniqueName: \"kubernetes.io/projected/eddaaabc-450b-4c69-9e29-35f554b01893-kube-api-access-ssf25\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921781 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/84fe49a3-baa7-45bf-aa7b-b92e55e61a1e-cert\") pod \"frr-k8s-webhook-server-6998585d5-wzm9n\" (UID: \"84fe49a3-baa7-45bf-aa7b-b92e55e61a1e\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921804 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljbmf\" (UniqueName: \"kubernetes.io/projected/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-kube-api-access-ljbmf\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921844 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-startup\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921874 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6b4j\" (UniqueName: \"kubernetes.io/projected/e40c967d-3d49-4b06-8789-9b1abe501e80-kube-api-access-n6b4j\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921896 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-reloader\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.921915 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.922119 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-metrics\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.922395 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-metrics\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.922693 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-frr-startup\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.930080 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-metrics-certs\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.932483 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/84fe49a3-baa7-45bf-aa7b-b92e55e61a1e-cert\") pod \"frr-k8s-webhook-server-6998585d5-wzm9n\" (UID: \"84fe49a3-baa7-45bf-aa7b-b92e55e61a1e\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.945801 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljbmf\" (UniqueName: \"kubernetes.io/projected/2719cd0e-b5e9-4df2-9165-b9bb102b1b70-kube-api-access-ljbmf\") pod \"frr-k8s-bkh26\" (UID: \"2719cd0e-b5e9-4df2-9165-b9bb102b1b70\") " pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:14 crc kubenswrapper[4735]: I1122 08:19:14.946211 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdkwc\" (UniqueName: \"kubernetes.io/projected/84fe49a3-baa7-45bf-aa7b-b92e55e61a1e-kube-api-access-kdkwc\") pod \"frr-k8s-webhook-server-6998585d5-wzm9n\" (UID: \"84fe49a3-baa7-45bf-aa7b-b92e55e61a1e\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.023435 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-metrics-certs\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.023811 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eddaaabc-450b-4c69-9e29-35f554b01893-metrics-certs\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.023953 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eddaaabc-450b-4c69-9e29-35f554b01893-cert\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.024111 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e40c967d-3d49-4b06-8789-9b1abe501e80-metallb-excludel2\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.024215 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssf25\" (UniqueName: \"kubernetes.io/projected/eddaaabc-450b-4c69-9e29-35f554b01893-kube-api-access-ssf25\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:15 crc kubenswrapper[4735]: E1122 08:19:15.023621 4735 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 22 08:19:15 crc kubenswrapper[4735]: E1122 08:19:15.024402 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-metrics-certs podName:e40c967d-3d49-4b06-8789-9b1abe501e80 nodeName:}" failed. No retries permitted until 2025-11-22 08:19:15.524379996 +0000 UTC m=+977.128718601 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-metrics-certs") pod "speaker-klgsh" (UID: "e40c967d-3d49-4b06-8789-9b1abe501e80") : secret "speaker-certs-secret" not found Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.024342 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6b4j\" (UniqueName: \"kubernetes.io/projected/e40c967d-3d49-4b06-8789-9b1abe501e80-kube-api-access-n6b4j\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.024525 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: E1122 08:19:15.024641 4735 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 08:19:15 crc kubenswrapper[4735]: E1122 08:19:15.024728 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist podName:e40c967d-3d49-4b06-8789-9b1abe501e80 nodeName:}" failed. No retries permitted until 2025-11-22 08:19:15.524692064 +0000 UTC m=+977.129030669 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist") pod "speaker-klgsh" (UID: "e40c967d-3d49-4b06-8789-9b1abe501e80") : secret "metallb-memberlist" not found Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.024813 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e40c967d-3d49-4b06-8789-9b1abe501e80-metallb-excludel2\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.026177 4735 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.027375 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eddaaabc-450b-4c69-9e29-35f554b01893-metrics-certs\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.036062 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.037654 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eddaaabc-450b-4c69-9e29-35f554b01893-cert\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.044793 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6b4j\" (UniqueName: \"kubernetes.io/projected/e40c967d-3d49-4b06-8789-9b1abe501e80-kube-api-access-n6b4j\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.047470 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.049361 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssf25\" (UniqueName: \"kubernetes.io/projected/eddaaabc-450b-4c69-9e29-35f554b01893-kube-api-access-ssf25\") pod \"controller-6c7b4b5f48-2gwq7\" (UID: \"eddaaabc-450b-4c69-9e29-35f554b01893\") " pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.170524 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.444853 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-2gwq7"] Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.534392 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-metrics-certs\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.534569 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: E1122 08:19:15.534741 4735 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 08:19:15 crc kubenswrapper[4735]: E1122 08:19:15.534798 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist podName:e40c967d-3d49-4b06-8789-9b1abe501e80 nodeName:}" failed. No retries permitted until 2025-11-22 08:19:16.534783989 +0000 UTC m=+978.139122594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist") pod "speaker-klgsh" (UID: "e40c967d-3d49-4b06-8789-9b1abe501e80") : secret "metallb-memberlist" not found Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.542200 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-metrics-certs\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.544695 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n"] Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.651750 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-2gwq7" event={"ID":"eddaaabc-450b-4c69-9e29-35f554b01893","Type":"ContainerStarted","Data":"26b4f758576b49d83696ef7d93392d0454313d693dd658a461b639a67e43b6e4"} Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.651793 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-2gwq7" event={"ID":"eddaaabc-450b-4c69-9e29-35f554b01893","Type":"ContainerStarted","Data":"4ecdf6b0e01c6142cbab28c3b4d3b7454952e1ef12f4220f0b5100d6a41d97b9"} Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.653206 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" event={"ID":"84fe49a3-baa7-45bf-aa7b-b92e55e61a1e","Type":"ContainerStarted","Data":"b066323aacd26522629e2594fbc21e9ea5a7c68e9c8c7811ee7c3ee3356b42dc"} Nov 22 08:19:15 crc kubenswrapper[4735]: I1122 08:19:15.654452 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerStarted","Data":"4161324d356287ab4ebfb81089be121f35b2a6f3413a99c128c2e80aae1b94d2"} Nov 22 08:19:16 crc kubenswrapper[4735]: I1122 08:19:16.557556 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:16 crc kubenswrapper[4735]: I1122 08:19:16.572917 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e40c967d-3d49-4b06-8789-9b1abe501e80-memberlist\") pod \"speaker-klgsh\" (UID: \"e40c967d-3d49-4b06-8789-9b1abe501e80\") " pod="metallb-system/speaker-klgsh" Nov 22 08:19:16 crc kubenswrapper[4735]: I1122 08:19:16.642446 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-klgsh" Nov 22 08:19:16 crc kubenswrapper[4735]: I1122 08:19:16.670351 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-2gwq7" event={"ID":"eddaaabc-450b-4c69-9e29-35f554b01893","Type":"ContainerStarted","Data":"68375babc4be0f9e0e27e1babe107c3ffa81ef648f7620df1e2f148781ec2b9a"} Nov 22 08:19:16 crc kubenswrapper[4735]: I1122 08:19:16.670966 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:16 crc kubenswrapper[4735]: I1122 08:19:16.694331 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-2gwq7" podStartSLOduration=2.694313824 podStartE2EDuration="2.694313824s" podCreationTimestamp="2025-11-22 08:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:16.686581887 +0000 UTC m=+978.290920492" watchObservedRunningTime="2025-11-22 08:19:16.694313824 +0000 UTC m=+978.298652429" Nov 22 08:19:17 crc kubenswrapper[4735]: I1122 08:19:17.684862 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-klgsh" event={"ID":"e40c967d-3d49-4b06-8789-9b1abe501e80","Type":"ContainerStarted","Data":"12af1fd5756613dd4a2dc97dd9973f4773e39a71d4204f45b4a219a701a17da5"} Nov 22 08:19:17 crc kubenswrapper[4735]: I1122 08:19:17.685174 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-klgsh" event={"ID":"e40c967d-3d49-4b06-8789-9b1abe501e80","Type":"ContainerStarted","Data":"84db5530305c7b8c53b6c1b2c311449f4fa97fe5e23cdb80bf73e03d57d672c6"} Nov 22 08:19:17 crc kubenswrapper[4735]: I1122 08:19:17.685188 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-klgsh" event={"ID":"e40c967d-3d49-4b06-8789-9b1abe501e80","Type":"ContainerStarted","Data":"ecbb5bb343b54e653139a5073bd0ae6a67bf973a4f29bdf1bd44948f3e2e9022"} Nov 22 08:19:17 crc kubenswrapper[4735]: I1122 08:19:17.685996 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-klgsh" Nov 22 08:19:17 crc kubenswrapper[4735]: I1122 08:19:17.732969 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-klgsh" podStartSLOduration=3.732946708 podStartE2EDuration="3.732946708s" podCreationTimestamp="2025-11-22 08:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:17.710704104 +0000 UTC m=+979.315042729" watchObservedRunningTime="2025-11-22 08:19:17.732946708 +0000 UTC m=+979.337285313" Nov 22 08:19:22 crc kubenswrapper[4735]: E1122 08:19:22.969191 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2719cd0e_b5e9_4df2_9165_b9bb102b1b70.slice/crio-0094d43346ed04f18b1c8c8a6931a45caa4c94597e1b4547986b1a8f0950f803.scope\": RecentStats: unable to find data in memory cache]" Nov 22 08:19:23 crc kubenswrapper[4735]: I1122 08:19:23.757332 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" event={"ID":"84fe49a3-baa7-45bf-aa7b-b92e55e61a1e","Type":"ContainerStarted","Data":"04ee9dd99d5b67178da73ab03f874b9d9933a6f60f2b76e931b798484e8d15a9"} Nov 22 08:19:23 crc kubenswrapper[4735]: I1122 08:19:23.757803 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:23 crc kubenswrapper[4735]: I1122 08:19:23.760451 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719cd0e-b5e9-4df2-9165-b9bb102b1b70" containerID="0094d43346ed04f18b1c8c8a6931a45caa4c94597e1b4547986b1a8f0950f803" exitCode=0 Nov 22 08:19:23 crc kubenswrapper[4735]: I1122 08:19:23.760663 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerDied","Data":"0094d43346ed04f18b1c8c8a6931a45caa4c94597e1b4547986b1a8f0950f803"} Nov 22 08:19:23 crc kubenswrapper[4735]: I1122 08:19:23.790817 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" podStartSLOduration=2.508014643 podStartE2EDuration="9.790797519s" podCreationTimestamp="2025-11-22 08:19:14 +0000 UTC" firstStartedPulling="2025-11-22 08:19:15.546134312 +0000 UTC m=+977.150472917" lastFinishedPulling="2025-11-22 08:19:22.828917188 +0000 UTC m=+984.433255793" observedRunningTime="2025-11-22 08:19:23.782700384 +0000 UTC m=+985.387038989" watchObservedRunningTime="2025-11-22 08:19:23.790797519 +0000 UTC m=+985.395136124" Nov 22 08:19:24 crc kubenswrapper[4735]: I1122 08:19:24.769497 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719cd0e-b5e9-4df2-9165-b9bb102b1b70" containerID="81ffa3223de281dd4bb1619e2314e270269168e4c44f09231383981964777451" exitCode=0 Nov 22 08:19:24 crc kubenswrapper[4735]: I1122 08:19:24.769541 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerDied","Data":"81ffa3223de281dd4bb1619e2314e270269168e4c44f09231383981964777451"} Nov 22 08:19:25 crc kubenswrapper[4735]: I1122 08:19:25.173991 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-2gwq7" Nov 22 08:19:25 crc kubenswrapper[4735]: I1122 08:19:25.779669 4735 generic.go:334] "Generic (PLEG): container finished" podID="2719cd0e-b5e9-4df2-9165-b9bb102b1b70" containerID="e29711e9456c66d73bf83f8b659a33f60f9ec82c86afa6be7f6ac81a9b9e0a4c" exitCode=0 Nov 22 08:19:25 crc kubenswrapper[4735]: I1122 08:19:25.779733 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerDied","Data":"e29711e9456c66d73bf83f8b659a33f60f9ec82c86afa6be7f6ac81a9b9e0a4c"} Nov 22 08:19:26 crc kubenswrapper[4735]: I1122 08:19:26.798802 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerStarted","Data":"b823d6efa2c18f65738f1fe034982722bc4e9342f80f8d5e87cc48fac17072a1"} Nov 22 08:19:26 crc kubenswrapper[4735]: I1122 08:19:26.799036 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerStarted","Data":"6fc18768e43ef32eb0b597cb0eb0cb3dc4aea67372cf456b2b5b77a4b7be4546"} Nov 22 08:19:26 crc kubenswrapper[4735]: I1122 08:19:26.799047 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerStarted","Data":"6c336ee63200e7b247d8b2b6fd0139ee19c9f9eba8c6561e4ae989342a5c86eb"} Nov 22 08:19:26 crc kubenswrapper[4735]: I1122 08:19:26.799055 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerStarted","Data":"40f81757baad3c1467578f7f160da7dbbc16f81cfcd3698a0aeed30801876faf"} Nov 22 08:19:26 crc kubenswrapper[4735]: I1122 08:19:26.799064 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerStarted","Data":"ba54979d48b27926e6ec58ec3b32dbfd8503f1d510152e83afc73d7a7e223e4c"} Nov 22 08:19:27 crc kubenswrapper[4735]: I1122 08:19:27.810484 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkh26" event={"ID":"2719cd0e-b5e9-4df2-9165-b9bb102b1b70","Type":"ContainerStarted","Data":"1a414fb728073a332affe99c1292e3b90b97ec6eefcde8bd16b3566201aac72f"} Nov 22 08:19:27 crc kubenswrapper[4735]: I1122 08:19:27.810682 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:27 crc kubenswrapper[4735]: I1122 08:19:27.834849 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-bkh26" podStartSLOduration=6.272997955 podStartE2EDuration="13.83482785s" podCreationTimestamp="2025-11-22 08:19:14 +0000 UTC" firstStartedPulling="2025-11-22 08:19:15.262263044 +0000 UTC m=+976.866601659" lastFinishedPulling="2025-11-22 08:19:22.824092949 +0000 UTC m=+984.428431554" observedRunningTime="2025-11-22 08:19:27.832410615 +0000 UTC m=+989.436749240" watchObservedRunningTime="2025-11-22 08:19:27.83482785 +0000 UTC m=+989.439166455" Nov 22 08:19:30 crc kubenswrapper[4735]: I1122 08:19:30.036866 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:30 crc kubenswrapper[4735]: I1122 08:19:30.071736 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:35 crc kubenswrapper[4735]: I1122 08:19:35.040129 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-bkh26" Nov 22 08:19:35 crc kubenswrapper[4735]: I1122 08:19:35.057687 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" Nov 22 08:19:36 crc kubenswrapper[4735]: I1122 08:19:36.648289 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-klgsh" Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.848167 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6txsc"] Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.850081 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6txsc" Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.852766 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.852887 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vcjbw" Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.853606 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.869610 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6txsc"] Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.871828 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9lk5\" (UniqueName: \"kubernetes.io/projected/93adb30e-1967-42b9-b39b-93486afbaf84-kube-api-access-h9lk5\") pod \"openstack-operator-index-6txsc\" (UID: \"93adb30e-1967-42b9-b39b-93486afbaf84\") " pod="openstack-operators/openstack-operator-index-6txsc" Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.973409 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9lk5\" (UniqueName: \"kubernetes.io/projected/93adb30e-1967-42b9-b39b-93486afbaf84-kube-api-access-h9lk5\") pod \"openstack-operator-index-6txsc\" (UID: \"93adb30e-1967-42b9-b39b-93486afbaf84\") " pod="openstack-operators/openstack-operator-index-6txsc" Nov 22 08:19:39 crc kubenswrapper[4735]: I1122 08:19:39.998034 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9lk5\" (UniqueName: \"kubernetes.io/projected/93adb30e-1967-42b9-b39b-93486afbaf84-kube-api-access-h9lk5\") pod \"openstack-operator-index-6txsc\" (UID: \"93adb30e-1967-42b9-b39b-93486afbaf84\") " pod="openstack-operators/openstack-operator-index-6txsc" Nov 22 08:19:40 crc kubenswrapper[4735]: I1122 08:19:40.205450 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6txsc" Nov 22 08:19:40 crc kubenswrapper[4735]: I1122 08:19:40.616068 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6txsc"] Nov 22 08:19:40 crc kubenswrapper[4735]: I1122 08:19:40.925330 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6txsc" event={"ID":"93adb30e-1967-42b9-b39b-93486afbaf84","Type":"ContainerStarted","Data":"51c7247c2bac98004386e38b0d55a034bb7eb54ad5a65c275efdf2c805d07e32"} Nov 22 08:19:43 crc kubenswrapper[4735]: I1122 08:19:43.579205 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6txsc"] Nov 22 08:19:43 crc kubenswrapper[4735]: I1122 08:19:43.948551 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6txsc" event={"ID":"93adb30e-1967-42b9-b39b-93486afbaf84","Type":"ContainerStarted","Data":"6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20"} Nov 22 08:19:43 crc kubenswrapper[4735]: I1122 08:19:43.965644 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-6txsc" podStartSLOduration=2.612796275 podStartE2EDuration="4.965624008s" podCreationTimestamp="2025-11-22 08:19:39 +0000 UTC" firstStartedPulling="2025-11-22 08:19:40.622622638 +0000 UTC m=+1002.226961243" lastFinishedPulling="2025-11-22 08:19:42.975450371 +0000 UTC m=+1004.579788976" observedRunningTime="2025-11-22 08:19:43.962992028 +0000 UTC m=+1005.567330633" watchObservedRunningTime="2025-11-22 08:19:43.965624008 +0000 UTC m=+1005.569962613" Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.181249 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-24jzf"] Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.182226 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.194953 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-24jzf"] Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.376802 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ds4w\" (UniqueName: \"kubernetes.io/projected/1eac1f05-c6ed-409f-802f-cb15adb08ef5-kube-api-access-2ds4w\") pod \"openstack-operator-index-24jzf\" (UID: \"1eac1f05-c6ed-409f-802f-cb15adb08ef5\") " pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.478620 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ds4w\" (UniqueName: \"kubernetes.io/projected/1eac1f05-c6ed-409f-802f-cb15adb08ef5-kube-api-access-2ds4w\") pod \"openstack-operator-index-24jzf\" (UID: \"1eac1f05-c6ed-409f-802f-cb15adb08ef5\") " pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.506323 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ds4w\" (UniqueName: \"kubernetes.io/projected/1eac1f05-c6ed-409f-802f-cb15adb08ef5-kube-api-access-2ds4w\") pod \"openstack-operator-index-24jzf\" (UID: \"1eac1f05-c6ed-409f-802f-cb15adb08ef5\") " pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.509916 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.958948 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-6txsc" podUID="93adb30e-1967-42b9-b39b-93486afbaf84" containerName="registry-server" containerID="cri-o://6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20" gracePeriod=2 Nov 22 08:19:44 crc kubenswrapper[4735]: I1122 08:19:44.984640 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-24jzf"] Nov 22 08:19:44 crc kubenswrapper[4735]: W1122 08:19:44.990617 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1eac1f05_c6ed_409f_802f_cb15adb08ef5.slice/crio-5a6d4c76f84664dbd3563f9c992974dfaf3bb0e6681fb399be352e5d031101d2 WatchSource:0}: Error finding container 5a6d4c76f84664dbd3563f9c992974dfaf3bb0e6681fb399be352e5d031101d2: Status 404 returned error can't find the container with id 5a6d4c76f84664dbd3563f9c992974dfaf3bb0e6681fb399be352e5d031101d2 Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.304239 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6txsc" Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.494715 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9lk5\" (UniqueName: \"kubernetes.io/projected/93adb30e-1967-42b9-b39b-93486afbaf84-kube-api-access-h9lk5\") pod \"93adb30e-1967-42b9-b39b-93486afbaf84\" (UID: \"93adb30e-1967-42b9-b39b-93486afbaf84\") " Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.501209 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93adb30e-1967-42b9-b39b-93486afbaf84-kube-api-access-h9lk5" (OuterVolumeSpecName: "kube-api-access-h9lk5") pod "93adb30e-1967-42b9-b39b-93486afbaf84" (UID: "93adb30e-1967-42b9-b39b-93486afbaf84"). InnerVolumeSpecName "kube-api-access-h9lk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.597808 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9lk5\" (UniqueName: \"kubernetes.io/projected/93adb30e-1967-42b9-b39b-93486afbaf84-kube-api-access-h9lk5\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.971428 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-24jzf" event={"ID":"1eac1f05-c6ed-409f-802f-cb15adb08ef5","Type":"ContainerStarted","Data":"303ba91711945e926a618ba05d14270efd6e74ec1f92d2f8f1e5aa149cd897e8"} Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.971556 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-24jzf" event={"ID":"1eac1f05-c6ed-409f-802f-cb15adb08ef5","Type":"ContainerStarted","Data":"5a6d4c76f84664dbd3563f9c992974dfaf3bb0e6681fb399be352e5d031101d2"} Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.974105 4735 generic.go:334] "Generic (PLEG): container finished" podID="93adb30e-1967-42b9-b39b-93486afbaf84" containerID="6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20" exitCode=0 Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.974149 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6txsc" Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.974182 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6txsc" event={"ID":"93adb30e-1967-42b9-b39b-93486afbaf84","Type":"ContainerDied","Data":"6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20"} Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.974228 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6txsc" event={"ID":"93adb30e-1967-42b9-b39b-93486afbaf84","Type":"ContainerDied","Data":"51c7247c2bac98004386e38b0d55a034bb7eb54ad5a65c275efdf2c805d07e32"} Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.974258 4735 scope.go:117] "RemoveContainer" containerID="6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20" Nov 22 08:19:45 crc kubenswrapper[4735]: I1122 08:19:45.991353 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-24jzf" podStartSLOduration=1.941252769 podStartE2EDuration="1.991333857s" podCreationTimestamp="2025-11-22 08:19:44 +0000 UTC" firstStartedPulling="2025-11-22 08:19:44.996967777 +0000 UTC m=+1006.601306392" lastFinishedPulling="2025-11-22 08:19:45.047048855 +0000 UTC m=+1006.651387480" observedRunningTime="2025-11-22 08:19:45.990056423 +0000 UTC m=+1007.594395068" watchObservedRunningTime="2025-11-22 08:19:45.991333857 +0000 UTC m=+1007.595672472" Nov 22 08:19:46 crc kubenswrapper[4735]: I1122 08:19:46.003194 4735 scope.go:117] "RemoveContainer" containerID="6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20" Nov 22 08:19:46 crc kubenswrapper[4735]: E1122 08:19:46.003959 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20\": container with ID starting with 6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20 not found: ID does not exist" containerID="6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20" Nov 22 08:19:46 crc kubenswrapper[4735]: I1122 08:19:46.003994 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20"} err="failed to get container status \"6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20\": rpc error: code = NotFound desc = could not find container \"6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20\": container with ID starting with 6eb14f05daf2c7b36fc695e375b719b3897f98bd48b43bc6ebb98e2375cf1f20 not found: ID does not exist" Nov 22 08:19:46 crc kubenswrapper[4735]: I1122 08:19:46.013991 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6txsc"] Nov 22 08:19:46 crc kubenswrapper[4735]: I1122 08:19:46.020670 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-6txsc"] Nov 22 08:19:47 crc kubenswrapper[4735]: I1122 08:19:47.281173 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93adb30e-1967-42b9-b39b-93486afbaf84" path="/var/lib/kubelet/pods/93adb30e-1967-42b9-b39b-93486afbaf84/volumes" Nov 22 08:19:54 crc kubenswrapper[4735]: I1122 08:19:54.511122 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:54 crc kubenswrapper[4735]: I1122 08:19:54.511549 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:54 crc kubenswrapper[4735]: I1122 08:19:54.540556 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.073231 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-24jzf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.818810 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf"] Nov 22 08:19:55 crc kubenswrapper[4735]: E1122 08:19:55.819242 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93adb30e-1967-42b9-b39b-93486afbaf84" containerName="registry-server" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.819262 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="93adb30e-1967-42b9-b39b-93486afbaf84" containerName="registry-server" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.819556 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="93adb30e-1967-42b9-b39b-93486afbaf84" containerName="registry-server" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.821436 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.824205 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dnh7g" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.827327 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf"] Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.894039 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-util\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.894347 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-bundle\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.894387 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh7wv\" (UniqueName: \"kubernetes.io/projected/783be8f3-61ed-4470-a034-f680ec16dbe1-kube-api-access-wh7wv\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.997261 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-util\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.997320 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-bundle\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.997362 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh7wv\" (UniqueName: \"kubernetes.io/projected/783be8f3-61ed-4470-a034-f680ec16dbe1-kube-api-access-wh7wv\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.998017 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-bundle\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:55 crc kubenswrapper[4735]: I1122 08:19:55.998324 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-util\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:56 crc kubenswrapper[4735]: I1122 08:19:56.016046 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh7wv\" (UniqueName: \"kubernetes.io/projected/783be8f3-61ed-4470-a034-f680ec16dbe1-kube-api-access-wh7wv\") pod \"2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:56 crc kubenswrapper[4735]: I1122 08:19:56.149203 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:19:56 crc kubenswrapper[4735]: I1122 08:19:56.615434 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf"] Nov 22 08:19:56 crc kubenswrapper[4735]: W1122 08:19:56.624817 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod783be8f3_61ed_4470_a034_f680ec16dbe1.slice/crio-d2789fce8809a7c3855661d41ac55823d9d9e81cccab8a57a04904ae2ebd8588 WatchSource:0}: Error finding container d2789fce8809a7c3855661d41ac55823d9d9e81cccab8a57a04904ae2ebd8588: Status 404 returned error can't find the container with id d2789fce8809a7c3855661d41ac55823d9d9e81cccab8a57a04904ae2ebd8588 Nov 22 08:19:57 crc kubenswrapper[4735]: I1122 08:19:57.083364 4735 generic.go:334] "Generic (PLEG): container finished" podID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerID="1788291644dc185c36bb17752276b8ff6b5942d12d75c9cd4b7f01daa07ddb80" exitCode=0 Nov 22 08:19:57 crc kubenswrapper[4735]: I1122 08:19:57.083420 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" event={"ID":"783be8f3-61ed-4470-a034-f680ec16dbe1","Type":"ContainerDied","Data":"1788291644dc185c36bb17752276b8ff6b5942d12d75c9cd4b7f01daa07ddb80"} Nov 22 08:19:57 crc kubenswrapper[4735]: I1122 08:19:57.083474 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" event={"ID":"783be8f3-61ed-4470-a034-f680ec16dbe1","Type":"ContainerStarted","Data":"d2789fce8809a7c3855661d41ac55823d9d9e81cccab8a57a04904ae2ebd8588"} Nov 22 08:19:58 crc kubenswrapper[4735]: I1122 08:19:58.093274 4735 generic.go:334] "Generic (PLEG): container finished" podID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerID="206cf194f0e2349cdb303514894f5a6870156b6a5011d1ee4131467b0eb4ad59" exitCode=0 Nov 22 08:19:58 crc kubenswrapper[4735]: I1122 08:19:58.093387 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" event={"ID":"783be8f3-61ed-4470-a034-f680ec16dbe1","Type":"ContainerDied","Data":"206cf194f0e2349cdb303514894f5a6870156b6a5011d1ee4131467b0eb4ad59"} Nov 22 08:19:59 crc kubenswrapper[4735]: I1122 08:19:59.107043 4735 generic.go:334] "Generic (PLEG): container finished" podID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerID="3e884b0f219707045e4a8435af7a3096038cb3e8934f9f7cfb0ff850159a5057" exitCode=0 Nov 22 08:19:59 crc kubenswrapper[4735]: I1122 08:19:59.107104 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" event={"ID":"783be8f3-61ed-4470-a034-f680ec16dbe1","Type":"ContainerDied","Data":"3e884b0f219707045e4a8435af7a3096038cb3e8934f9f7cfb0ff850159a5057"} Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.413625 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.576785 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-util\") pod \"783be8f3-61ed-4470-a034-f680ec16dbe1\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.576966 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh7wv\" (UniqueName: \"kubernetes.io/projected/783be8f3-61ed-4470-a034-f680ec16dbe1-kube-api-access-wh7wv\") pod \"783be8f3-61ed-4470-a034-f680ec16dbe1\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.577044 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-bundle\") pod \"783be8f3-61ed-4470-a034-f680ec16dbe1\" (UID: \"783be8f3-61ed-4470-a034-f680ec16dbe1\") " Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.578282 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-bundle" (OuterVolumeSpecName: "bundle") pod "783be8f3-61ed-4470-a034-f680ec16dbe1" (UID: "783be8f3-61ed-4470-a034-f680ec16dbe1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.583282 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/783be8f3-61ed-4470-a034-f680ec16dbe1-kube-api-access-wh7wv" (OuterVolumeSpecName: "kube-api-access-wh7wv") pod "783be8f3-61ed-4470-a034-f680ec16dbe1" (UID: "783be8f3-61ed-4470-a034-f680ec16dbe1"). InnerVolumeSpecName "kube-api-access-wh7wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.593157 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-util" (OuterVolumeSpecName: "util") pod "783be8f3-61ed-4470-a034-f680ec16dbe1" (UID: "783be8f3-61ed-4470-a034-f680ec16dbe1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.678632 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh7wv\" (UniqueName: \"kubernetes.io/projected/783be8f3-61ed-4470-a034-f680ec16dbe1-kube-api-access-wh7wv\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.678692 4735 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:00 crc kubenswrapper[4735]: I1122 08:20:00.678705 4735 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/783be8f3-61ed-4470-a034-f680ec16dbe1-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:01 crc kubenswrapper[4735]: I1122 08:20:01.128352 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" event={"ID":"783be8f3-61ed-4470-a034-f680ec16dbe1","Type":"ContainerDied","Data":"d2789fce8809a7c3855661d41ac55823d9d9e81cccab8a57a04904ae2ebd8588"} Nov 22 08:20:01 crc kubenswrapper[4735]: I1122 08:20:01.128398 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2789fce8809a7c3855661d41ac55823d9d9e81cccab8a57a04904ae2ebd8588" Nov 22 08:20:01 crc kubenswrapper[4735]: I1122 08:20:01.128754 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.384757 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8"] Nov 22 08:20:08 crc kubenswrapper[4735]: E1122 08:20:08.385663 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerName="pull" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.385678 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerName="pull" Nov 22 08:20:08 crc kubenswrapper[4735]: E1122 08:20:08.385703 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerName="extract" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.385710 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerName="extract" Nov 22 08:20:08 crc kubenswrapper[4735]: E1122 08:20:08.385722 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerName="util" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.385731 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerName="util" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.385927 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="783be8f3-61ed-4470-a034-f680ec16dbe1" containerName="extract" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.387202 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.390722 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-9b54v" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.415335 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8"] Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.432586 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmbwx\" (UniqueName: \"kubernetes.io/projected/7ba21bad-f0f3-4e1e-add2-43b2ddfae492-kube-api-access-tmbwx\") pod \"openstack-operator-controller-operator-7475f94df-sfjx8\" (UID: \"7ba21bad-f0f3-4e1e-add2-43b2ddfae492\") " pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.534416 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmbwx\" (UniqueName: \"kubernetes.io/projected/7ba21bad-f0f3-4e1e-add2-43b2ddfae492-kube-api-access-tmbwx\") pod \"openstack-operator-controller-operator-7475f94df-sfjx8\" (UID: \"7ba21bad-f0f3-4e1e-add2-43b2ddfae492\") " pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.558036 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmbwx\" (UniqueName: \"kubernetes.io/projected/7ba21bad-f0f3-4e1e-add2-43b2ddfae492-kube-api-access-tmbwx\") pod \"openstack-operator-controller-operator-7475f94df-sfjx8\" (UID: \"7ba21bad-f0f3-4e1e-add2-43b2ddfae492\") " pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" Nov 22 08:20:08 crc kubenswrapper[4735]: I1122 08:20:08.706675 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" Nov 22 08:20:09 crc kubenswrapper[4735]: I1122 08:20:09.157388 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8"] Nov 22 08:20:09 crc kubenswrapper[4735]: I1122 08:20:09.188423 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" event={"ID":"7ba21bad-f0f3-4e1e-add2-43b2ddfae492","Type":"ContainerStarted","Data":"0841622bd3c7e28ba94a9612d0641526a962a27d897914642b561048dd66b8e6"} Nov 22 08:20:14 crc kubenswrapper[4735]: I1122 08:20:14.232637 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" event={"ID":"7ba21bad-f0f3-4e1e-add2-43b2ddfae492","Type":"ContainerStarted","Data":"e73b64e4724863c431e697d83ab58089cf33a21e14928bbab6e1cf03294f8304"} Nov 22 08:20:17 crc kubenswrapper[4735]: I1122 08:20:17.255318 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" event={"ID":"7ba21bad-f0f3-4e1e-add2-43b2ddfae492","Type":"ContainerStarted","Data":"c7c25e5da1e0fa3f74ff196c0baf47096dc570d100fc2aa18fad2ad45ab314cb"} Nov 22 08:20:17 crc kubenswrapper[4735]: I1122 08:20:17.255977 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" Nov 22 08:20:17 crc kubenswrapper[4735]: I1122 08:20:17.300631 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" podStartSLOduration=1.75336921 podStartE2EDuration="9.300608953s" podCreationTimestamp="2025-11-22 08:20:08 +0000 UTC" firstStartedPulling="2025-11-22 08:20:09.166298876 +0000 UTC m=+1030.770637481" lastFinishedPulling="2025-11-22 08:20:16.713538619 +0000 UTC m=+1038.317877224" observedRunningTime="2025-11-22 08:20:17.29418931 +0000 UTC m=+1038.898527925" watchObservedRunningTime="2025-11-22 08:20:17.300608953 +0000 UTC m=+1038.904947558" Nov 22 08:20:18 crc kubenswrapper[4735]: I1122 08:20:18.265346 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7475f94df-sfjx8" Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.938023 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp"] Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.940823 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.943255 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb"] Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.944866 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.945624 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-8wh2g" Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.946222 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-c8qk7" Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.959870 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn"] Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.961872 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.967395 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-72977" Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.983396 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp"] Nov 22 08:20:35 crc kubenswrapper[4735]: I1122 08:20:35.995024 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.002779 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.017492 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-zrv98"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.019774 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.029840 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-ws4fl" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.061728 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-zrv98"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.072959 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.076496 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.078348 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5fnq5" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.081539 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.090891 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.092454 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.094655 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-fnn54" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.106370 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.111897 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxvl4\" (UniqueName: \"kubernetes.io/projected/2038726f-ef9a-47d2-837e-1b81f72c2d39-kube-api-access-mxvl4\") pod \"glance-operator-controller-manager-7969689c84-zrv98\" (UID: \"2038726f-ef9a-47d2-837e-1b81f72c2d39\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.112001 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnvtt\" (UniqueName: \"kubernetes.io/projected/2dd266b7-8f4d-46b3-b8e7-3511b5404f8d-kube-api-access-lnvtt\") pod \"cinder-operator-controller-manager-6498cbf48f-26vdp\" (UID: \"2dd266b7-8f4d-46b3-b8e7-3511b5404f8d\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.112058 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xqkw\" (UniqueName: \"kubernetes.io/projected/2c87c2e3-b68e-4f75-a7cf-2c752e7286e8-kube-api-access-4xqkw\") pod \"designate-operator-controller-manager-767ccfd65f-j5xsn\" (UID: \"2c87c2e3-b68e-4f75-a7cf-2c752e7286e8\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.112089 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxshx\" (UniqueName: \"kubernetes.io/projected/5765b990-2731-42c7-bb14-7b149e91555f-kube-api-access-qxshx\") pod \"barbican-operator-controller-manager-75fb479bcc-dmhxb\" (UID: \"5765b990-2731-42c7-bb14-7b149e91555f\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.128811 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.130624 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.135804 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.135999 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vmdh9" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.147526 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-t625d"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.149178 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.151992 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hqzb7" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.157859 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.159660 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.167868 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-w7f4t" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.168042 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.169316 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.170981 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hnznb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.179535 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.205025 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-t625d"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.228115 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfhf9\" (UniqueName: \"kubernetes.io/projected/2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb-kube-api-access-zfhf9\") pod \"keystone-operator-controller-manager-7454b96578-t625d\" (UID: \"2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.228406 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxvl4\" (UniqueName: \"kubernetes.io/projected/2038726f-ef9a-47d2-837e-1b81f72c2d39-kube-api-access-mxvl4\") pod \"glance-operator-controller-manager-7969689c84-zrv98\" (UID: \"2038726f-ef9a-47d2-837e-1b81f72c2d39\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.228549 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k559\" (UniqueName: \"kubernetes.io/projected/1c5cb76f-8964-4245-86e5-56b619e06d1b-kube-api-access-8k559\") pod \"heat-operator-controller-manager-56f54d6746-gkpjb\" (UID: \"1c5cb76f-8964-4245-86e5-56b619e06d1b\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.228655 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnvtt\" (UniqueName: \"kubernetes.io/projected/2dd266b7-8f4d-46b3-b8e7-3511b5404f8d-kube-api-access-lnvtt\") pod \"cinder-operator-controller-manager-6498cbf48f-26vdp\" (UID: \"2dd266b7-8f4d-46b3-b8e7-3511b5404f8d\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.228760 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xqkw\" (UniqueName: \"kubernetes.io/projected/2c87c2e3-b68e-4f75-a7cf-2c752e7286e8-kube-api-access-4xqkw\") pod \"designate-operator-controller-manager-767ccfd65f-j5xsn\" (UID: \"2c87c2e3-b68e-4f75-a7cf-2c752e7286e8\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.228846 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxshx\" (UniqueName: \"kubernetes.io/projected/5765b990-2731-42c7-bb14-7b149e91555f-kube-api-access-qxshx\") pod \"barbican-operator-controller-manager-75fb479bcc-dmhxb\" (UID: \"5765b990-2731-42c7-bb14-7b149e91555f\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.228935 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgqbq\" (UniqueName: \"kubernetes.io/projected/c3ac6a85-2f1e-4906-ba7d-c9415f523fa8-kube-api-access-mgqbq\") pod \"horizon-operator-controller-manager-598f69df5d-42kvc\" (UID: \"c3ac6a85-2f1e-4906-ba7d-c9415f523fa8\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.254935 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.262263 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnvtt\" (UniqueName: \"kubernetes.io/projected/2dd266b7-8f4d-46b3-b8e7-3511b5404f8d-kube-api-access-lnvtt\") pod \"cinder-operator-controller-manager-6498cbf48f-26vdp\" (UID: \"2dd266b7-8f4d-46b3-b8e7-3511b5404f8d\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.262558 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxshx\" (UniqueName: \"kubernetes.io/projected/5765b990-2731-42c7-bb14-7b149e91555f-kube-api-access-qxshx\") pod \"barbican-operator-controller-manager-75fb479bcc-dmhxb\" (UID: \"5765b990-2731-42c7-bb14-7b149e91555f\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.279231 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.280157 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxvl4\" (UniqueName: \"kubernetes.io/projected/2038726f-ef9a-47d2-837e-1b81f72c2d39-kube-api-access-mxvl4\") pod \"glance-operator-controller-manager-7969689c84-zrv98\" (UID: \"2038726f-ef9a-47d2-837e-1b81f72c2d39\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.284565 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.286941 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.318318 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-bqn4n" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.318773 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xqkw\" (UniqueName: \"kubernetes.io/projected/2c87c2e3-b68e-4f75-a7cf-2c752e7286e8-kube-api-access-4xqkw\") pod \"designate-operator-controller-manager-767ccfd65f-j5xsn\" (UID: \"2c87c2e3-b68e-4f75-a7cf-2c752e7286e8\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.319230 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.330926 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k559\" (UniqueName: \"kubernetes.io/projected/1c5cb76f-8964-4245-86e5-56b619e06d1b-kube-api-access-8k559\") pod \"heat-operator-controller-manager-56f54d6746-gkpjb\" (UID: \"1c5cb76f-8964-4245-86e5-56b619e06d1b\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.331014 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgqbq\" (UniqueName: \"kubernetes.io/projected/c3ac6a85-2f1e-4906-ba7d-c9415f523fa8-kube-api-access-mgqbq\") pod \"horizon-operator-controller-manager-598f69df5d-42kvc\" (UID: \"c3ac6a85-2f1e-4906-ba7d-c9415f523fa8\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.331041 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r82g7\" (UniqueName: \"kubernetes.io/projected/61871ccb-bdb6-4222-88d9-d34c22a5797c-kube-api-access-r82g7\") pod \"ironic-operator-controller-manager-99b499f4-pwlw2\" (UID: \"61871ccb-bdb6-4222-88d9-d34c22a5797c\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.331064 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10f9fa68-fe17-4af9-8e47-fd534c8a4446-cert\") pod \"infra-operator-controller-manager-7875d8bb94-tqj8q\" (UID: \"10f9fa68-fe17-4af9-8e47-fd534c8a4446\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.331100 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g64hz\" (UniqueName: \"kubernetes.io/projected/fc7fded7-7eee-4227-8569-85d07ea4ecd6-kube-api-access-g64hz\") pod \"manila-operator-controller-manager-58f887965d-nd4wh\" (UID: \"fc7fded7-7eee-4227-8569-85d07ea4ecd6\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.331117 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfhf9\" (UniqueName: \"kubernetes.io/projected/2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb-kube-api-access-zfhf9\") pod \"keystone-operator-controller-manager-7454b96578-t625d\" (UID: \"2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.334583 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74dtp\" (UniqueName: \"kubernetes.io/projected/10f9fa68-fe17-4af9-8e47-fd534c8a4446-kube-api-access-74dtp\") pod \"infra-operator-controller-manager-7875d8bb94-tqj8q\" (UID: \"10f9fa68-fe17-4af9-8e47-fd534c8a4446\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.349396 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.356124 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.361635 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.363735 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.366676 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-x52jn" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.391687 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k559\" (UniqueName: \"kubernetes.io/projected/1c5cb76f-8964-4245-86e5-56b619e06d1b-kube-api-access-8k559\") pod \"heat-operator-controller-manager-56f54d6746-gkpjb\" (UID: \"1c5cb76f-8964-4245-86e5-56b619e06d1b\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.395350 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgqbq\" (UniqueName: \"kubernetes.io/projected/c3ac6a85-2f1e-4906-ba7d-c9415f523fa8-kube-api-access-mgqbq\") pod \"horizon-operator-controller-manager-598f69df5d-42kvc\" (UID: \"c3ac6a85-2f1e-4906-ba7d-c9415f523fa8\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.396145 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfhf9\" (UniqueName: \"kubernetes.io/projected/2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb-kube-api-access-zfhf9\") pod \"keystone-operator-controller-manager-7454b96578-t625d\" (UID: \"2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.416034 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.430085 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.435888 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g64hz\" (UniqueName: \"kubernetes.io/projected/fc7fded7-7eee-4227-8569-85d07ea4ecd6-kube-api-access-g64hz\") pod \"manila-operator-controller-manager-58f887965d-nd4wh\" (UID: \"fc7fded7-7eee-4227-8569-85d07ea4ecd6\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.436112 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74dtp\" (UniqueName: \"kubernetes.io/projected/10f9fa68-fe17-4af9-8e47-fd534c8a4446-kube-api-access-74dtp\") pod \"infra-operator-controller-manager-7875d8bb94-tqj8q\" (UID: \"10f9fa68-fe17-4af9-8e47-fd534c8a4446\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.436365 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlrdk\" (UniqueName: \"kubernetes.io/projected/3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c-kube-api-access-mlrdk\") pod \"mariadb-operator-controller-manager-54b5986bb8-vgsgb\" (UID: \"3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.441271 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r82g7\" (UniqueName: \"kubernetes.io/projected/61871ccb-bdb6-4222-88d9-d34c22a5797c-kube-api-access-r82g7\") pod \"ironic-operator-controller-manager-99b499f4-pwlw2\" (UID: \"61871ccb-bdb6-4222-88d9-d34c22a5797c\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.442549 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10f9fa68-fe17-4af9-8e47-fd534c8a4446-cert\") pod \"infra-operator-controller-manager-7875d8bb94-tqj8q\" (UID: \"10f9fa68-fe17-4af9-8e47-fd534c8a4446\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.437502 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.439574 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv"] Nov 22 08:20:36 crc kubenswrapper[4735]: E1122 08:20:36.444871 4735 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 22 08:20:36 crc kubenswrapper[4735]: E1122 08:20:36.445082 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10f9fa68-fe17-4af9-8e47-fd534c8a4446-cert podName:10f9fa68-fe17-4af9-8e47-fd534c8a4446 nodeName:}" failed. No retries permitted until 2025-11-22 08:20:36.945043027 +0000 UTC m=+1058.549381632 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/10f9fa68-fe17-4af9-8e47-fd534c8a4446-cert") pod "infra-operator-controller-manager-7875d8bb94-tqj8q" (UID: "10f9fa68-fe17-4af9-8e47-fd534c8a4446") : secret "infra-operator-webhook-server-cert" not found Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.449406 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.454363 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.461902 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9j5pr" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.462261 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74dtp\" (UniqueName: \"kubernetes.io/projected/10f9fa68-fe17-4af9-8e47-fd534c8a4446-kube-api-access-74dtp\") pod \"infra-operator-controller-manager-7875d8bb94-tqj8q\" (UID: \"10f9fa68-fe17-4af9-8e47-fd534c8a4446\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.468876 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.470585 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.470826 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r82g7\" (UniqueName: \"kubernetes.io/projected/61871ccb-bdb6-4222-88d9-d34c22a5797c-kube-api-access-r82g7\") pod \"ironic-operator-controller-manager-99b499f4-pwlw2\" (UID: \"61871ccb-bdb6-4222-88d9-d34c22a5797c\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.473947 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-p4c4q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.480746 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g64hz\" (UniqueName: \"kubernetes.io/projected/fc7fded7-7eee-4227-8569-85d07ea4ecd6-kube-api-access-g64hz\") pod \"manila-operator-controller-manager-58f887965d-nd4wh\" (UID: \"fc7fded7-7eee-4227-8569-85d07ea4ecd6\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.494001 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.503567 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.546208 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.546908 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.549372 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlrdk\" (UniqueName: \"kubernetes.io/projected/3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c-kube-api-access-mlrdk\") pod \"mariadb-operator-controller-manager-54b5986bb8-vgsgb\" (UID: \"3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.552852 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltxvx\" (UniqueName: \"kubernetes.io/projected/e204b082-d231-48c6-af16-f4ee9d2d7f9c-kube-api-access-ltxvx\") pod \"octavia-operator-controller-manager-54cfbf4c7d-2kf8g\" (UID: \"e204b082-d231-48c6-af16-f4ee9d2d7f9c\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.552940 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnx49\" (UniqueName: \"kubernetes.io/projected/9124e032-f3c0-46fa-bd17-26ed40bfd57e-kube-api-access-nnx49\") pod \"neutron-operator-controller-manager-78bd47f458-xl2mv\" (UID: \"9124e032-f3c0-46fa-bd17-26ed40bfd57e\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.553257 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fz8d\" (UniqueName: \"kubernetes.io/projected/150b26f8-b5cd-4828-96cc-6e37000c24e7-kube-api-access-9fz8d\") pod \"nova-operator-controller-manager-cfbb9c588-b47vs\" (UID: \"150b26f8-b5cd-4828-96cc-6e37000c24e7\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.568420 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.597192 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlrdk\" (UniqueName: \"kubernetes.io/projected/3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c-kube-api-access-mlrdk\") pod \"mariadb-operator-controller-manager-54b5986bb8-vgsgb\" (UID: \"3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.606362 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.654440 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.655012 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fz8d\" (UniqueName: \"kubernetes.io/projected/150b26f8-b5cd-4828-96cc-6e37000c24e7-kube-api-access-9fz8d\") pod \"nova-operator-controller-manager-cfbb9c588-b47vs\" (UID: \"150b26f8-b5cd-4828-96cc-6e37000c24e7\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.655107 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltxvx\" (UniqueName: \"kubernetes.io/projected/e204b082-d231-48c6-af16-f4ee9d2d7f9c-kube-api-access-ltxvx\") pod \"octavia-operator-controller-manager-54cfbf4c7d-2kf8g\" (UID: \"e204b082-d231-48c6-af16-f4ee9d2d7f9c\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.655130 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnx49\" (UniqueName: \"kubernetes.io/projected/9124e032-f3c0-46fa-bd17-26ed40bfd57e-kube-api-access-nnx49\") pod \"neutron-operator-controller-manager-78bd47f458-xl2mv\" (UID: \"9124e032-f3c0-46fa-bd17-26ed40bfd57e\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.656393 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.669003 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.669253 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-7nn6r" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.675629 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.677275 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.680020 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnx49\" (UniqueName: \"kubernetes.io/projected/9124e032-f3c0-46fa-bd17-26ed40bfd57e-kube-api-access-nnx49\") pod \"neutron-operator-controller-manager-78bd47f458-xl2mv\" (UID: \"9124e032-f3c0-46fa-bd17-26ed40bfd57e\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.682591 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-slsmc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.682834 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fz8d\" (UniqueName: \"kubernetes.io/projected/150b26f8-b5cd-4828-96cc-6e37000c24e7-kube-api-access-9fz8d\") pod \"nova-operator-controller-manager-cfbb9c588-b47vs\" (UID: \"150b26f8-b5cd-4828-96cc-6e37000c24e7\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.685011 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltxvx\" (UniqueName: \"kubernetes.io/projected/e204b082-d231-48c6-af16-f4ee9d2d7f9c-kube-api-access-ltxvx\") pod \"octavia-operator-controller-manager-54cfbf4c7d-2kf8g\" (UID: \"e204b082-d231-48c6-af16-f4ee9d2d7f9c\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.704439 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.705996 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.709048 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-khj7t" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.756395 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.756934 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hz62\" (UniqueName: \"kubernetes.io/projected/2171ac01-b987-4062-9521-88c800930330-kube-api-access-7hz62\") pod \"placement-operator-controller-manager-5b797b8dff-vndrw\" (UID: \"2171ac01-b987-4062-9521-88c800930330\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.757004 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vblnr\" (UniqueName: \"kubernetes.io/projected/990c8bbe-2866-4568-8e30-1d9f016f0736-kube-api-access-vblnr\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc\" (UID: \"990c8bbe-2866-4568-8e30-1d9f016f0736\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.757055 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/990c8bbe-2866-4568-8e30-1d9f016f0736-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc\" (UID: \"990c8bbe-2866-4568-8e30-1d9f016f0736\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.757080 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw6fv\" (UniqueName: \"kubernetes.io/projected/85c63b21-5764-45af-98bc-47b1be4be90e-kube-api-access-qw6fv\") pod \"ovn-operator-controller-manager-54fc5f65b7-qs98r\" (UID: \"85c63b21-5764-45af-98bc-47b1be4be90e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.771930 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.778621 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.780327 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.785213 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wfgq5" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.786143 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.809259 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.818106 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.853615 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.854896 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.858505 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq9fg\" (UniqueName: \"kubernetes.io/projected/69fe7050-a836-427e-ba06-7f016e64b24b-kube-api-access-jq9fg\") pod \"swift-operator-controller-manager-d656998f4-x7w9w\" (UID: \"69fe7050-a836-427e-ba06-7f016e64b24b\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.858554 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vblnr\" (UniqueName: \"kubernetes.io/projected/990c8bbe-2866-4568-8e30-1d9f016f0736-kube-api-access-vblnr\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc\" (UID: \"990c8bbe-2866-4568-8e30-1d9f016f0736\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.858616 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/990c8bbe-2866-4568-8e30-1d9f016f0736-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc\" (UID: \"990c8bbe-2866-4568-8e30-1d9f016f0736\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.858646 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw6fv\" (UniqueName: \"kubernetes.io/projected/85c63b21-5764-45af-98bc-47b1be4be90e-kube-api-access-qw6fv\") pod \"ovn-operator-controller-manager-54fc5f65b7-qs98r\" (UID: \"85c63b21-5764-45af-98bc-47b1be4be90e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.858694 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s28gx\" (UniqueName: \"kubernetes.io/projected/ee3c788a-51f9-4808-996b-e854ba26a776-kube-api-access-s28gx\") pod \"telemetry-operator-controller-manager-7d76d46c6f-shjbx\" (UID: \"ee3c788a-51f9-4808-996b-e854ba26a776\") " pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.858742 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hz62\" (UniqueName: \"kubernetes.io/projected/2171ac01-b987-4062-9521-88c800930330-kube-api-access-7hz62\") pod \"placement-operator-controller-manager-5b797b8dff-vndrw\" (UID: \"2171ac01-b987-4062-9521-88c800930330\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" Nov 22 08:20:36 crc kubenswrapper[4735]: E1122 08:20:36.859240 4735 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 22 08:20:36 crc kubenswrapper[4735]: E1122 08:20:36.859278 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/990c8bbe-2866-4568-8e30-1d9f016f0736-cert podName:990c8bbe-2866-4568-8e30-1d9f016f0736 nodeName:}" failed. No retries permitted until 2025-11-22 08:20:37.359266783 +0000 UTC m=+1058.963605378 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/990c8bbe-2866-4568-8e30-1d9f016f0736-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" (UID: "990c8bbe-2866-4568-8e30-1d9f016f0736") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.864250 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-bs9ht" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.864401 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.870251 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.908182 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.913843 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vblnr\" (UniqueName: \"kubernetes.io/projected/990c8bbe-2866-4568-8e30-1d9f016f0736-kube-api-access-vblnr\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc\" (UID: \"990c8bbe-2866-4568-8e30-1d9f016f0736\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.928397 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw6fv\" (UniqueName: \"kubernetes.io/projected/85c63b21-5764-45af-98bc-47b1be4be90e-kube-api-access-qw6fv\") pod \"ovn-operator-controller-manager-54fc5f65b7-qs98r\" (UID: \"85c63b21-5764-45af-98bc-47b1be4be90e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.928492 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.934037 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hz62\" (UniqueName: \"kubernetes.io/projected/2171ac01-b987-4062-9521-88c800930330-kube-api-access-7hz62\") pod \"placement-operator-controller-manager-5b797b8dff-vndrw\" (UID: \"2171ac01-b987-4062-9521-88c800930330\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.969224 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz"] Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.970679 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.972425 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10f9fa68-fe17-4af9-8e47-fd534c8a4446-cert\") pod \"infra-operator-controller-manager-7875d8bb94-tqj8q\" (UID: \"10f9fa68-fe17-4af9-8e47-fd534c8a4446\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.972482 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s28gx\" (UniqueName: \"kubernetes.io/projected/ee3c788a-51f9-4808-996b-e854ba26a776-kube-api-access-s28gx\") pod \"telemetry-operator-controller-manager-7d76d46c6f-shjbx\" (UID: \"ee3c788a-51f9-4808-996b-e854ba26a776\") " pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.972584 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq9fg\" (UniqueName: \"kubernetes.io/projected/69fe7050-a836-427e-ba06-7f016e64b24b-kube-api-access-jq9fg\") pod \"swift-operator-controller-manager-d656998f4-x7w9w\" (UID: \"69fe7050-a836-427e-ba06-7f016e64b24b\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.977160 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10f9fa68-fe17-4af9-8e47-fd534c8a4446-cert\") pod \"infra-operator-controller-manager-7875d8bb94-tqj8q\" (UID: \"10f9fa68-fe17-4af9-8e47-fd534c8a4446\") " pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.978934 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-r2mxj" Nov 22 08:20:36 crc kubenswrapper[4735]: I1122 08:20:36.993315 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.030760 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s28gx\" (UniqueName: \"kubernetes.io/projected/ee3c788a-51f9-4808-996b-e854ba26a776-kube-api-access-s28gx\") pod \"telemetry-operator-controller-manager-7d76d46c6f-shjbx\" (UID: \"ee3c788a-51f9-4808-996b-e854ba26a776\") " pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.042808 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq9fg\" (UniqueName: \"kubernetes.io/projected/69fe7050-a836-427e-ba06-7f016e64b24b-kube-api-access-jq9fg\") pod \"swift-operator-controller-manager-d656998f4-x7w9w\" (UID: \"69fe7050-a836-427e-ba06-7f016e64b24b\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.058758 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.060705 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.062136 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.065593 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.069748 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-rl9h5" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.088093 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.097865 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h68n6\" (UniqueName: \"kubernetes.io/projected/fce6a4d4-f6fa-44bf-8528-7de867925bbb-kube-api-access-h68n6\") pod \"watcher-operator-controller-manager-8c6448b9f-cn8lg\" (UID: \"fce6a4d4-f6fa-44bf-8528-7de867925bbb\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.098040 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2slcc\" (UniqueName: \"kubernetes.io/projected/fb88fa45-ef38-49bc-b842-157bef4cd0e5-kube-api-access-2slcc\") pod \"test-operator-controller-manager-b4c496f69-dcgnz\" (UID: \"fb88fa45-ef38-49bc-b842-157bef4cd0e5\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.103789 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.126149 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.136749 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.152563 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.155079 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5mb92" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.155258 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.169813 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.202432 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s858p\" (UniqueName: \"kubernetes.io/projected/b83d5866-e478-4d10-967a-488ef0435473-kube-api-access-s858p\") pod \"openstack-operator-controller-manager-76b989774d-x7gnj\" (UID: \"b83d5866-e478-4d10-967a-488ef0435473\") " pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.208922 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b83d5866-e478-4d10-967a-488ef0435473-cert\") pod \"openstack-operator-controller-manager-76b989774d-x7gnj\" (UID: \"b83d5866-e478-4d10-967a-488ef0435473\") " pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.208991 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h68n6\" (UniqueName: \"kubernetes.io/projected/fce6a4d4-f6fa-44bf-8528-7de867925bbb-kube-api-access-h68n6\") pod \"watcher-operator-controller-manager-8c6448b9f-cn8lg\" (UID: \"fce6a4d4-f6fa-44bf-8528-7de867925bbb\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.209216 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2slcc\" (UniqueName: \"kubernetes.io/projected/fb88fa45-ef38-49bc-b842-157bef4cd0e5-kube-api-access-2slcc\") pod \"test-operator-controller-manager-b4c496f69-dcgnz\" (UID: \"fb88fa45-ef38-49bc-b842-157bef4cd0e5\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.215556 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.216578 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.221802 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-tw4dv" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.225064 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.229609 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.243554 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2slcc\" (UniqueName: \"kubernetes.io/projected/fb88fa45-ef38-49bc-b842-157bef4cd0e5-kube-api-access-2slcc\") pod \"test-operator-controller-manager-b4c496f69-dcgnz\" (UID: \"fb88fa45-ef38-49bc-b842-157bef4cd0e5\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.254986 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h68n6\" (UniqueName: \"kubernetes.io/projected/fce6a4d4-f6fa-44bf-8528-7de867925bbb-kube-api-access-h68n6\") pod \"watcher-operator-controller-manager-8c6448b9f-cn8lg\" (UID: \"fce6a4d4-f6fa-44bf-8528-7de867925bbb\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.305015 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.321424 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b83d5866-e478-4d10-967a-488ef0435473-cert\") pod \"openstack-operator-controller-manager-76b989774d-x7gnj\" (UID: \"b83d5866-e478-4d10-967a-488ef0435473\") " pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.321561 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l944f\" (UniqueName: \"kubernetes.io/projected/11f148e7-7f11-42cb-ae64-e5714bf56295-kube-api-access-l944f\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-285tt\" (UID: \"11f148e7-7f11-42cb-ae64-e5714bf56295\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.321640 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s858p\" (UniqueName: \"kubernetes.io/projected/b83d5866-e478-4d10-967a-488ef0435473-kube-api-access-s858p\") pod \"openstack-operator-controller-manager-76b989774d-x7gnj\" (UID: \"b83d5866-e478-4d10-967a-488ef0435473\") " pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:37 crc kubenswrapper[4735]: E1122 08:20:37.321958 4735 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 22 08:20:37 crc kubenswrapper[4735]: E1122 08:20:37.322032 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b83d5866-e478-4d10-967a-488ef0435473-cert podName:b83d5866-e478-4d10-967a-488ef0435473 nodeName:}" failed. No retries permitted until 2025-11-22 08:20:37.822011784 +0000 UTC m=+1059.426350389 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b83d5866-e478-4d10-967a-488ef0435473-cert") pod "openstack-operator-controller-manager-76b989774d-x7gnj" (UID: "b83d5866-e478-4d10-967a-488ef0435473") : secret "webhook-server-cert" not found Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.330889 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.350445 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s858p\" (UniqueName: \"kubernetes.io/projected/b83d5866-e478-4d10-967a-488ef0435473-kube-api-access-s858p\") pod \"openstack-operator-controller-manager-76b989774d-x7gnj\" (UID: \"b83d5866-e478-4d10-967a-488ef0435473\") " pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.362968 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.368531 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-zrv98"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.402741 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.417439 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" event={"ID":"2dd266b7-8f4d-46b3-b8e7-3511b5404f8d","Type":"ContainerStarted","Data":"f67dd1644540323d9dbe69bd0c3960052b4d0a187e3356d79a85cca46f544031"} Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.424723 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/990c8bbe-2866-4568-8e30-1d9f016f0736-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc\" (UID: \"990c8bbe-2866-4568-8e30-1d9f016f0736\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.424766 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l944f\" (UniqueName: \"kubernetes.io/projected/11f148e7-7f11-42cb-ae64-e5714bf56295-kube-api-access-l944f\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-285tt\" (UID: \"11f148e7-7f11-42cb-ae64-e5714bf56295\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.430273 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/990c8bbe-2866-4568-8e30-1d9f016f0736-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc\" (UID: \"990c8bbe-2866-4568-8e30-1d9f016f0736\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.448978 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l944f\" (UniqueName: \"kubernetes.io/projected/11f148e7-7f11-42cb-ae64-e5714bf56295-kube-api-access-l944f\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-285tt\" (UID: \"11f148e7-7f11-42cb-ae64-e5714bf56295\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.590070 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.624271 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.627743 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb"] Nov 22 08:20:37 crc kubenswrapper[4735]: W1122 08:20:37.648481 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c5cb76f_8964_4245_86e5_56b619e06d1b.slice/crio-7a09cefb1f2547cfecfda6769419083f1ece225d36b1cb43dd91a6400b1f599c WatchSource:0}: Error finding container 7a09cefb1f2547cfecfda6769419083f1ece225d36b1cb43dd91a6400b1f599c: Status 404 returned error can't find the container with id 7a09cefb1f2547cfecfda6769419083f1ece225d36b1cb43dd91a6400b1f599c Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.834391 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b83d5866-e478-4d10-967a-488ef0435473-cert\") pod \"openstack-operator-controller-manager-76b989774d-x7gnj\" (UID: \"b83d5866-e478-4d10-967a-488ef0435473\") " pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.840493 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b83d5866-e478-4d10-967a-488ef0435473-cert\") pod \"openstack-operator-controller-manager-76b989774d-x7gnj\" (UID: \"b83d5866-e478-4d10-967a-488ef0435473\") " pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.964688 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc"] Nov 22 08:20:37 crc kubenswrapper[4735]: W1122 08:20:37.973680 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2270ad73_e96d_4e1e_bd2c_e6fb254a1bdb.slice/crio-5d839b4afa77e9168a0b38c5e4cc639f6cb43f85c218b741823120eaa9046fea WatchSource:0}: Error finding container 5d839b4afa77e9168a0b38c5e4cc639f6cb43f85c218b741823120eaa9046fea: Status 404 returned error can't find the container with id 5d839b4afa77e9168a0b38c5e4cc639f6cb43f85c218b741823120eaa9046fea Nov 22 08:20:37 crc kubenswrapper[4735]: W1122 08:20:37.974809 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3ac6a85_2f1e_4906_ba7d_c9415f523fa8.slice/crio-a357b39b469c90413e6ce11da137ed95e564827e46f5853717e5f0e0db46a29f WatchSource:0}: Error finding container a357b39b469c90413e6ce11da137ed95e564827e46f5853717e5f0e0db46a29f: Status 404 returned error can't find the container with id a357b39b469c90413e6ce11da137ed95e564827e46f5853717e5f0e0db46a29f Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.984409 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-t625d"] Nov 22 08:20:37 crc kubenswrapper[4735]: I1122 08:20:37.993567 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.085219 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.329887 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.384504 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.394882 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.404646 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.415922 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.447200 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.448707 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" event={"ID":"3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c","Type":"ContainerStarted","Data":"56c31581b344bcef18e9d557265b6bbbfd7fe454b86363d7f9799364776636cd"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.461640 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" event={"ID":"2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb","Type":"ContainerStarted","Data":"5d839b4afa77e9168a0b38c5e4cc639f6cb43f85c218b741823120eaa9046fea"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.463522 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" event={"ID":"2c87c2e3-b68e-4f75-a7cf-2c752e7286e8","Type":"ContainerStarted","Data":"b420d46d0a266208219b426d1d20c09e9eadf1ec206da1cd6a978ec3bb94a10a"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.478986 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" event={"ID":"1c5cb76f-8964-4245-86e5-56b619e06d1b","Type":"ContainerStarted","Data":"7a09cefb1f2547cfecfda6769419083f1ece225d36b1cb43dd91a6400b1f599c"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.481740 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" event={"ID":"5765b990-2731-42c7-bb14-7b149e91555f","Type":"ContainerStarted","Data":"656da38b14c9a2122efaa16548c90b4fc3f7fc1b9676aedf62a96f469a515608"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.482807 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" event={"ID":"10f9fa68-fe17-4af9-8e47-fd534c8a4446","Type":"ContainerStarted","Data":"1cc96b449277af493e1d346f21bc1920a830736e8a98b296a2fcbd8653af45d7"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.489451 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" event={"ID":"c3ac6a85-2f1e-4906-ba7d-c9415f523fa8","Type":"ContainerStarted","Data":"a357b39b469c90413e6ce11da137ed95e564827e46f5853717e5f0e0db46a29f"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.490692 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" event={"ID":"fc7fded7-7eee-4227-8569-85d07ea4ecd6","Type":"ContainerStarted","Data":"b8c8a738589b775181c204ec596b929690c5feae8e59c6382ca3b6cd76007a38"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.493279 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" event={"ID":"2038726f-ef9a-47d2-837e-1b81f72c2d39","Type":"ContainerStarted","Data":"65a858e54579f2ba78b072c11df41d6681aa5a26b3284e5292ee545524ad6210"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.494821 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" event={"ID":"150b26f8-b5cd-4828-96cc-6e37000c24e7","Type":"ContainerStarted","Data":"d72fcb2d194776e2b573e56d1e82adc939117923ad42bc299fa62ac4f8c9404a"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.497003 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" event={"ID":"e204b082-d231-48c6-af16-f4ee9d2d7f9c","Type":"ContainerStarted","Data":"759cc594d16b6c05c3dc63558f2c06bd6b03fae5e9c9113c29165b9ff4c0948e"} Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.716033 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.726531 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.735747 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.742480 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv"] Nov 22 08:20:38 crc kubenswrapper[4735]: W1122 08:20:38.744814 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb88fa45_ef38_49bc_b842_157bef4cd0e5.slice/crio-d9d8dad0b1851e9d705eb1141ab86f53606ebcf888ba88739d985507a0d14b68 WatchSource:0}: Error finding container d9d8dad0b1851e9d705eb1141ab86f53606ebcf888ba88739d985507a0d14b68: Status 404 returned error can't find the container with id d9d8dad0b1851e9d705eb1141ab86f53606ebcf888ba88739d985507a0d14b68 Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.750758 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz"] Nov 22 08:20:38 crc kubenswrapper[4735]: I1122 08:20:38.757942 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w"] Nov 22 08:20:38 crc kubenswrapper[4735]: W1122 08:20:38.764761 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9124e032_f3c0_46fa_bd17_26ed40bfd57e.slice/crio-b646cbe4082e756b7b2cc669e8ea10c3ffd39fc105d0ae9f195f690773c7e409 WatchSource:0}: Error finding container b646cbe4082e756b7b2cc669e8ea10c3ffd39fc105d0ae9f195f690773c7e409: Status 404 returned error can't find the container with id b646cbe4082e756b7b2cc669e8ea10c3ffd39fc105d0ae9f195f690773c7e409 Nov 22 08:20:38 crc kubenswrapper[4735]: W1122 08:20:38.766955 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69fe7050_a836_427e_ba06_7f016e64b24b.slice/crio-34630c58c4e551791353ac9c22e80048c3a1fa736297025575bec3d763bdb2c9 WatchSource:0}: Error finding container 34630c58c4e551791353ac9c22e80048c3a1fa736297025575bec3d763bdb2c9: Status 404 returned error can't find the container with id 34630c58c4e551791353ac9c22e80048c3a1fa736297025575bec3d763bdb2c9 Nov 22 08:20:38 crc kubenswrapper[4735]: E1122 08:20:38.770952 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nnx49,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-78bd47f458-xl2mv_openstack-operators(9124e032-f3c0-46fa-bd17-26ed40bfd57e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:20:38 crc kubenswrapper[4735]: E1122 08:20:38.771735 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jq9fg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d656998f4-x7w9w_openstack-operators(69fe7050-a836-427e-ba06-7f016e64b24b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:20:38 crc kubenswrapper[4735]: E1122 08:20:38.981306 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" podUID="9124e032-f3c0-46fa-bd17-26ed40bfd57e" Nov 22 08:20:39 crc kubenswrapper[4735]: E1122 08:20:39.047405 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" podUID="69fe7050-a836-427e-ba06-7f016e64b24b" Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.114538 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw"] Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.131688 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt"] Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.139536 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc"] Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.163753 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj"] Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.548630 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" event={"ID":"ee3c788a-51f9-4808-996b-e854ba26a776","Type":"ContainerStarted","Data":"3c8dd60ac33f5ab1d93084007a43afd5f92288ba22ee79da3661690c5811e153"} Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.568679 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" event={"ID":"85c63b21-5764-45af-98bc-47b1be4be90e","Type":"ContainerStarted","Data":"3845e1bf5275397662794baa24c268597fe9fc1685bd9113e82b45c53a92b10d"} Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.602478 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" event={"ID":"69fe7050-a836-427e-ba06-7f016e64b24b","Type":"ContainerStarted","Data":"5983668d60e443eedbdc914cdde3ce222409c8536d28bf5d6eda14d84f046c02"} Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.602524 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" event={"ID":"69fe7050-a836-427e-ba06-7f016e64b24b","Type":"ContainerStarted","Data":"34630c58c4e551791353ac9c22e80048c3a1fa736297025575bec3d763bdb2c9"} Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.619680 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" event={"ID":"fce6a4d4-f6fa-44bf-8528-7de867925bbb","Type":"ContainerStarted","Data":"f5ca63e67ff8d6406c120f72fa94ae422fe7ae4793eedd0aa001a9956f6e7401"} Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.634324 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" event={"ID":"61871ccb-bdb6-4222-88d9-d34c22a5797c","Type":"ContainerStarted","Data":"dd91a5bd33816b37287623bf1252ee8bff0b7f0d2dc6463b9cf5e389a53ec6b9"} Nov 22 08:20:39 crc kubenswrapper[4735]: E1122 08:20:39.638670 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" podUID="69fe7050-a836-427e-ba06-7f016e64b24b" Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.643932 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" event={"ID":"9124e032-f3c0-46fa-bd17-26ed40bfd57e","Type":"ContainerStarted","Data":"4cd3de5417a1ee0f33b66678cfb1e981119cabaaa888b0c652965a7f8561e9f5"} Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.643971 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" event={"ID":"9124e032-f3c0-46fa-bd17-26ed40bfd57e","Type":"ContainerStarted","Data":"b646cbe4082e756b7b2cc669e8ea10c3ffd39fc105d0ae9f195f690773c7e409"} Nov 22 08:20:39 crc kubenswrapper[4735]: E1122 08:20:39.645676 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" podUID="9124e032-f3c0-46fa-bd17-26ed40bfd57e" Nov 22 08:20:39 crc kubenswrapper[4735]: I1122 08:20:39.653656 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" event={"ID":"fb88fa45-ef38-49bc-b842-157bef4cd0e5","Type":"ContainerStarted","Data":"d9d8dad0b1851e9d705eb1141ab86f53606ebcf888ba88739d985507a0d14b68"} Nov 22 08:20:40 crc kubenswrapper[4735]: E1122 08:20:40.673027 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" podUID="69fe7050-a836-427e-ba06-7f016e64b24b" Nov 22 08:20:40 crc kubenswrapper[4735]: E1122 08:20:40.674193 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" podUID="9124e032-f3c0-46fa-bd17-26ed40bfd57e" Nov 22 08:20:46 crc kubenswrapper[4735]: I1122 08:20:46.131588 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:20:46 crc kubenswrapper[4735]: I1122 08:20:46.132164 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:21:01 crc kubenswrapper[4735]: E1122 08:21:01.398857 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a" Nov 22 08:21:01 crc kubenswrapper[4735]: E1122 08:21:01.399978 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zfhf9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7454b96578-t625d_openstack-operators(2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:02 crc kubenswrapper[4735]: W1122 08:21:02.350282 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb83d5866_e478_4d10_967a_488ef0435473.slice/crio-bb035534f204bf8a82230e19bb2feb9ede708947100a77052dafbb1c277e79c1 WatchSource:0}: Error finding container bb035534f204bf8a82230e19bb2feb9ede708947100a77052dafbb1c277e79c1: Status 404 returned error can't find the container with id bb035534f204bf8a82230e19bb2feb9ede708947100a77052dafbb1c277e79c1 Nov 22 08:21:02 crc kubenswrapper[4735]: W1122 08:21:02.353706 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11f148e7_7f11_42cb_ae64_e5714bf56295.slice/crio-ab1e505b7e1bf255a53fba6a9d6f45c2b6221e47b7a7ba2bc8e6b2b17b43d505 WatchSource:0}: Error finding container ab1e505b7e1bf255a53fba6a9d6f45c2b6221e47b7a7ba2bc8e6b2b17b43d505: Status 404 returned error can't find the container with id ab1e505b7e1bf255a53fba6a9d6f45c2b6221e47b7a7ba2bc8e6b2b17b43d505 Nov 22 08:21:02 crc kubenswrapper[4735]: I1122 08:21:02.875499 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" event={"ID":"b83d5866-e478-4d10-967a-488ef0435473","Type":"ContainerStarted","Data":"bb035534f204bf8a82230e19bb2feb9ede708947100a77052dafbb1c277e79c1"} Nov 22 08:21:02 crc kubenswrapper[4735]: I1122 08:21:02.876708 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" event={"ID":"11f148e7-7f11-42cb-ae64-e5714bf56295","Type":"ContainerStarted","Data":"ab1e505b7e1bf255a53fba6a9d6f45c2b6221e47b7a7ba2bc8e6b2b17b43d505"} Nov 22 08:21:03 crc kubenswrapper[4735]: E1122 08:21:03.160319 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a" Nov 22 08:21:03 crc kubenswrapper[4735]: E1122 08:21:03.160556 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g64hz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-58f887965d-nd4wh_openstack-operators(fc7fded7-7eee-4227-8569-85d07ea4ecd6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:04 crc kubenswrapper[4735]: I1122 08:21:04.895819 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" event={"ID":"2171ac01-b987-4062-9521-88c800930330","Type":"ContainerStarted","Data":"23ca0ae512707e5d6967a56192c46ebb87a81472e7432c41b827f50088da350c"} Nov 22 08:21:04 crc kubenswrapper[4735]: I1122 08:21:04.897578 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" event={"ID":"990c8bbe-2866-4568-8e30-1d9f016f0736","Type":"ContainerStarted","Data":"9573f5c0912aedb36418dad270be53a4f38c5076346889d6697935e7975db318"} Nov 22 08:21:05 crc kubenswrapper[4735]: E1122 08:21:05.289224 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96" Nov 22 08:21:05 crc kubenswrapper[4735]: E1122 08:21:05.290028 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8k559,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-56f54d6746-gkpjb_openstack-operators(1c5cb76f-8964-4245-86e5-56b619e06d1b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:07 crc kubenswrapper[4735]: E1122 08:21:07.937944 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:d38faa9070da05487afdaa9e261ad39274c2ed862daf42efa460a040431f1991" Nov 22 08:21:07 crc kubenswrapper[4735]: E1122 08:21:07.938413 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:d38faa9070da05487afdaa9e261ad39274c2ed862daf42efa460a040431f1991,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mxvl4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-7969689c84-zrv98_openstack-operators(2038726f-ef9a-47d2-837e-1b81f72c2d39): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:08 crc kubenswrapper[4735]: E1122 08:21:08.743781 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9" Nov 22 08:21:08 crc kubenswrapper[4735]: E1122 08:21:08.743979 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mgqbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-598f69df5d-42kvc_openstack-operators(c3ac6a85-2f1e-4906-ba7d-c9415f523fa8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:09 crc kubenswrapper[4735]: E1122 08:21:09.693689 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:f0688f6a55b7b548aaafd5c2c4f0749a43e7ea447c62a24e8b35257c5d8ba17f" Nov 22 08:21:09 crc kubenswrapper[4735]: E1122 08:21:09.693870 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:f0688f6a55b7b548aaafd5c2c4f0749a43e7ea447c62a24e8b35257c5d8ba17f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-74dtp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-7875d8bb94-tqj8q_openstack-operators(10f9fa68-fe17-4af9-8e47-fd534c8a4446): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:15 crc kubenswrapper[4735]: E1122 08:21:15.310480 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377" Nov 22 08:21:15 crc kubenswrapper[4735]: E1122 08:21:15.311315 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r82g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-99b499f4-pwlw2_openstack-operators(61871ccb-bdb6-4222-88d9-d34c22a5797c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:16 crc kubenswrapper[4735]: I1122 08:21:16.131772 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:21:16 crc kubenswrapper[4735]: I1122 08:21:16.131840 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:21:17 crc kubenswrapper[4735]: E1122 08:21:17.893393 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13" Nov 22 08:21:17 crc kubenswrapper[4735]: E1122 08:21:17.893906 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ltxvx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-54cfbf4c7d-2kf8g_openstack-operators(e204b082-d231-48c6-af16-f4ee9d2d7f9c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:18 crc kubenswrapper[4735]: E1122 08:21:18.338052 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7" Nov 22 08:21:18 crc kubenswrapper[4735]: E1122 08:21:18.338268 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qxshx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-75fb479bcc-dmhxb_openstack-operators(5765b990-2731-42c7-bb14-7b149e91555f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:19 crc kubenswrapper[4735]: E1122 08:21:19.386906 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7" Nov 22 08:21:19 crc kubenswrapper[4735]: E1122 08:21:19.387331 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9fz8d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-cfbb9c588-b47vs_openstack-operators(150b26f8-b5cd-4828-96cc-6e37000c24e7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:19 crc kubenswrapper[4735]: E1122 08:21:19.833367 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d" Nov 22 08:21:19 crc kubenswrapper[4735]: E1122 08:21:19.833576 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2slcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-dcgnz_openstack-operators(fb88fa45-ef38-49bc-b842-157bef4cd0e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:20 crc kubenswrapper[4735]: E1122 08:21:20.269564 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c" Nov 22 08:21:20 crc kubenswrapper[4735]: E1122 08:21:20.269783 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7hz62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b797b8dff-vndrw_openstack-operators(2171ac01-b987-4062-9521-88c800930330): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:22 crc kubenswrapper[4735]: E1122 08:21:22.483877 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.51:5001/openstack-k8s-operators/telemetry-operator:0a8fdb2f5e6b6d1c474c1888450357b75d025375" Nov 22 08:21:22 crc kubenswrapper[4735]: E1122 08:21:22.484428 4735 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.51:5001/openstack-k8s-operators/telemetry-operator:0a8fdb2f5e6b6d1c474c1888450357b75d025375" Nov 22 08:21:22 crc kubenswrapper[4735]: E1122 08:21:22.484597 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.129.56.51:5001/openstack-k8s-operators/telemetry-operator:0a8fdb2f5e6b6d1c474c1888450357b75d025375,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s28gx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7d76d46c6f-shjbx_openstack-operators(ee3c788a-51f9-4808-996b-e854ba26a776): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:23 crc kubenswrapper[4735]: E1122 08:21:23.734039 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04" Nov 22 08:21:23 crc kubenswrapper[4735]: E1122 08:21:23.734534 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mlrdk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-54b5986bb8-vgsgb_openstack-operators(3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:24 crc kubenswrapper[4735]: E1122 08:21:24.168430 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6" Nov 22 08:21:24 crc kubenswrapper[4735]: E1122 08:21:24.168694 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nnx49,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-78bd47f458-xl2mv_openstack-operators(9124e032-f3c0-46fa-bd17-26ed40bfd57e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:24 crc kubenswrapper[4735]: E1122 08:21:24.169946 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" podUID="9124e032-f3c0-46fa-bd17-26ed40bfd57e" Nov 22 08:21:25 crc kubenswrapper[4735]: E1122 08:21:25.056984 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:c6405d94e56b40ef669729216ab4b9c441f34bb280902efa2940038c076b560f" Nov 22 08:21:25 crc kubenswrapper[4735]: E1122 08:21:25.057527 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:c6405d94e56b40ef669729216ab4b9c441f34bb280902efa2940038c076b560f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4xqkw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-767ccfd65f-j5xsn_openstack-operators(2c87c2e3-b68e-4f75-a7cf-2c752e7286e8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:25 crc kubenswrapper[4735]: E1122 08:21:25.479081 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd" Nov 22 08:21:25 crc kubenswrapper[4735]: E1122 08:21:25.479644 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vblnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc_openstack-operators(990c8bbe-2866-4568-8e30-1d9f016f0736): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:27 crc kubenswrapper[4735]: E1122 08:21:27.118931 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 22 08:21:27 crc kubenswrapper[4735]: E1122 08:21:27.119185 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h68n6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-cn8lg_openstack-operators(fce6a4d4-f6fa-44bf-8528-7de867925bbb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:27 crc kubenswrapper[4735]: E1122 08:21:27.531067 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 22 08:21:27 crc kubenswrapper[4735]: E1122 08:21:27.531492 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l944f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-285tt_openstack-operators(11f148e7-7f11-42cb-ae64-e5714bf56295): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:21:27 crc kubenswrapper[4735]: E1122 08:21:27.533511 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" podUID="11f148e7-7f11-42cb-ae64-e5714bf56295" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.107952 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" podUID="fc7fded7-7eee-4227-8569-85d07ea4ecd6" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.111228 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" podUID="11f148e7-7f11-42cb-ae64-e5714bf56295" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.169017 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" podUID="3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.182852 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" podUID="10f9fa68-fe17-4af9-8e47-fd534c8a4446" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.184849 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" podUID="5765b990-2731-42c7-bb14-7b149e91555f" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.200959 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" podUID="150b26f8-b5cd-4828-96cc-6e37000c24e7" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.212134 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" podUID="c3ac6a85-2f1e-4906-ba7d-c9415f523fa8" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.263746 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" podUID="2038726f-ef9a-47d2-837e-1b81f72c2d39" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.263834 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" podUID="2171ac01-b987-4062-9521-88c800930330" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.337736 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" podUID="fb88fa45-ef38-49bc-b842-157bef4cd0e5" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.424571 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" podUID="2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.466041 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" podUID="61871ccb-bdb6-4222-88d9-d34c22a5797c" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.556219 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" podUID="1c5cb76f-8964-4245-86e5-56b619e06d1b" Nov 22 08:21:28 crc kubenswrapper[4735]: E1122 08:21:28.922359 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" podUID="e204b082-d231-48c6-af16-f4ee9d2d7f9c" Nov 22 08:21:29 crc kubenswrapper[4735]: E1122 08:21:29.033853 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" podUID="2c87c2e3-b68e-4f75-a7cf-2c752e7286e8" Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.119203 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" event={"ID":"b83d5866-e478-4d10-967a-488ef0435473","Type":"ContainerStarted","Data":"2f2e4f0eec207b798943b13fedcaef73327ffb152647bd43066f53c028d16822"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.121029 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" event={"ID":"85c63b21-5764-45af-98bc-47b1be4be90e","Type":"ContainerStarted","Data":"73b55f395a82c720da936313295ec9b563a7f1ff66831ce84d5634ce7d78549d"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.122556 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" event={"ID":"fc7fded7-7eee-4227-8569-85d07ea4ecd6","Type":"ContainerStarted","Data":"fd468e5645dd314e413a600eafa0b9f5f7f0e53d23bf8c6b37aeb5c58ac903a3"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.124783 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" event={"ID":"2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb","Type":"ContainerStarted","Data":"f39aa7888152115d9b4caf98c1b46e5e6414f6868cb92a031b363351c2d784cc"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.126443 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" event={"ID":"2dd266b7-8f4d-46b3-b8e7-3511b5404f8d","Type":"ContainerStarted","Data":"cc3c3973e597c24081db91f5fe4ee1169fc325f5af9f98d1adc6fc27e4a20d61"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.128335 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" event={"ID":"1c5cb76f-8964-4245-86e5-56b619e06d1b","Type":"ContainerStarted","Data":"308e23e3df31c66372819f941b84d79c07527dd69320e5fa748b93ae840013ba"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.130881 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" event={"ID":"150b26f8-b5cd-4828-96cc-6e37000c24e7","Type":"ContainerStarted","Data":"7dbbd1480b8a2264db082052434d024ca7aa2218cd18d6ede1013996c94dacd5"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.133552 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" event={"ID":"5765b990-2731-42c7-bb14-7b149e91555f","Type":"ContainerStarted","Data":"c73d76be2d0f68cc0857d3053b6db35822f19ebae491ecc9bf09104a26a1f282"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.135028 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" event={"ID":"3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c","Type":"ContainerStarted","Data":"3cf358b9939cf23c50c9d638b89b42d3d78b72e8ec2dcbd996b89becbee31cc5"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.137032 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" event={"ID":"fce6a4d4-f6fa-44bf-8528-7de867925bbb","Type":"ContainerStarted","Data":"8bfcbecca0cfccd4070765da6d389203b60f0476162ce6ae51e3104cd1d7af78"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.138443 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" event={"ID":"61871ccb-bdb6-4222-88d9-d34c22a5797c","Type":"ContainerStarted","Data":"a41aac2993b4e21f7bcbb9aad2810a5164073bb73d2163356ea764cfa6cf764f"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.139634 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" event={"ID":"990c8bbe-2866-4568-8e30-1d9f016f0736","Type":"ContainerStarted","Data":"91aa6f15e0a2b72d6eb4934e14e4f35db4ca16280863d23f74846b50c19ff06b"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.141219 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" event={"ID":"e204b082-d231-48c6-af16-f4ee9d2d7f9c","Type":"ContainerStarted","Data":"efba498268042cb4991f85417a79ff0af2a894ee46a0784fa0f968404817d14f"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.143762 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" event={"ID":"fb88fa45-ef38-49bc-b842-157bef4cd0e5","Type":"ContainerStarted","Data":"f5a8d55d9c6d203fc344546ded2d8d682a5065c3e43bc93a3de267452fa4ab77"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.145452 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" event={"ID":"69fe7050-a836-427e-ba06-7f016e64b24b","Type":"ContainerStarted","Data":"77757d26cd5c3c74cb64791abd2048aea909e881d4d897f88380ce6de1503d7c"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.148453 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" event={"ID":"2038726f-ef9a-47d2-837e-1b81f72c2d39","Type":"ContainerStarted","Data":"850c2679d3ae45e2fc2f69f34a0b1cc0a302169d40027b7469bcc12c638df8e8"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.149642 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" event={"ID":"2c87c2e3-b68e-4f75-a7cf-2c752e7286e8","Type":"ContainerStarted","Data":"dfbd67c068012fedd4d122bc364b0a77906b09ead471414650254e4decd5af8f"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.150703 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" event={"ID":"2171ac01-b987-4062-9521-88c800930330","Type":"ContainerStarted","Data":"95679691b328aa6ca584ff841b728bc588c00fad5a5571a4d8c1e7e291f94184"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.159383 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" event={"ID":"10f9fa68-fe17-4af9-8e47-fd534c8a4446","Type":"ContainerStarted","Data":"bf924ced0af3e022ff1b05e5144ea95549538368ba6127a12e092624d6bacca8"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.161706 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" event={"ID":"c3ac6a85-2f1e-4906-ba7d-c9415f523fa8","Type":"ContainerStarted","Data":"1b7f8364c4ea581c4be021550a917f30e99c763f01265ea13e685808bfcb01e3"} Nov 22 08:21:29 crc kubenswrapper[4735]: I1122 08:21:29.167534 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" event={"ID":"ee3c788a-51f9-4808-996b-e854ba26a776","Type":"ContainerStarted","Data":"72319f1f024240a73a882a5270e5c740ede81a02beb8e01c6f0c1858a5b5dbaf"} Nov 22 08:21:30 crc kubenswrapper[4735]: E1122 08:21:30.670533 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" podUID="990c8bbe-2866-4568-8e30-1d9f016f0736" Nov 22 08:21:30 crc kubenswrapper[4735]: E1122 08:21:30.847712 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" podUID="ee3c788a-51f9-4808-996b-e854ba26a776" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.208816 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" podUID="fce6a4d4-f6fa-44bf-8528-7de867925bbb" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.229167 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" podUID="990c8bbe-2866-4568-8e30-1d9f016f0736" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.229395 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:c6405d94e56b40ef669729216ab4b9c441f34bb280902efa2940038c076b560f\\\"\"" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" podUID="2c87c2e3-b68e-4f75-a7cf-2c752e7286e8" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.229890 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.51:5001/openstack-k8s-operators/telemetry-operator:0a8fdb2f5e6b6d1c474c1888450357b75d025375\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" podUID="ee3c788a-51f9-4808-996b-e854ba26a776" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.230315 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" podUID="3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.230630 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qxshx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-75fb479bcc-dmhxb_openstack-operators(5765b990-2731-42c7-bb14-7b149e91555f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.230689 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9fz8d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-cfbb9c588-b47vs_openstack-operators(150b26f8-b5cd-4828-96cc-6e37000c24e7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.231880 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" podUID="150b26f8-b5cd-4828-96cc-6e37000c24e7" Nov 22 08:21:31 crc kubenswrapper[4735]: E1122 08:21:31.232057 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" podUID="5765b990-2731-42c7-bb14-7b149e91555f" Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.209411 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" event={"ID":"2dd266b7-8f4d-46b3-b8e7-3511b5404f8d","Type":"ContainerStarted","Data":"c91d923a99de5ee9478f9df9c95c31da412e1e47d278449f94f4f3ea4997c484"} Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.209753 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.217316 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" event={"ID":"b83d5866-e478-4d10-967a-488ef0435473","Type":"ContainerStarted","Data":"a164e5f62d14559be85e0a9dcc0f8590a9072f4fcbc0f7e15fa9e4c4c82526e2"} Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.217415 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.222603 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" event={"ID":"85c63b21-5764-45af-98bc-47b1be4be90e","Type":"ContainerStarted","Data":"adff59b877db116bf7199be10f3cdee1abc0d482a228753d198b581a069dad7c"} Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.223387 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.237028 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" podStartSLOduration=12.022714609 podStartE2EDuration="57.237002495s" podCreationTimestamp="2025-11-22 08:20:35 +0000 UTC" firstStartedPulling="2025-11-22 08:20:37.174843032 +0000 UTC m=+1058.779181637" lastFinishedPulling="2025-11-22 08:21:22.389130918 +0000 UTC m=+1103.993469523" observedRunningTime="2025-11-22 08:21:32.22890106 +0000 UTC m=+1113.833239665" watchObservedRunningTime="2025-11-22 08:21:32.237002495 +0000 UTC m=+1113.841341110" Nov 22 08:21:32 crc kubenswrapper[4735]: E1122 08:21:32.247616 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" podUID="fce6a4d4-f6fa-44bf-8528-7de867925bbb" Nov 22 08:21:32 crc kubenswrapper[4735]: E1122 08:21:32.247694 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:c6405d94e56b40ef669729216ab4b9c441f34bb280902efa2940038c076b560f\\\"\"" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" podUID="2c87c2e3-b68e-4f75-a7cf-2c752e7286e8" Nov 22 08:21:32 crc kubenswrapper[4735]: E1122 08:21:32.247712 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" podUID="5765b990-2731-42c7-bb14-7b149e91555f" Nov 22 08:21:32 crc kubenswrapper[4735]: E1122 08:21:32.247696 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" podUID="3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c" Nov 22 08:21:32 crc kubenswrapper[4735]: E1122 08:21:32.247763 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" podUID="150b26f8-b5cd-4828-96cc-6e37000c24e7" Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.269850 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" podStartSLOduration=8.428891581 podStartE2EDuration="56.269828077s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.771655961 +0000 UTC m=+1060.375994566" lastFinishedPulling="2025-11-22 08:21:26.612592437 +0000 UTC m=+1108.216931062" observedRunningTime="2025-11-22 08:21:32.264042187 +0000 UTC m=+1113.868380892" watchObservedRunningTime="2025-11-22 08:21:32.269828077 +0000 UTC m=+1113.874166682" Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.305362 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" podStartSLOduration=56.305324624 podStartE2EDuration="56.305324624s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:21:32.28898424 +0000 UTC m=+1113.893322845" watchObservedRunningTime="2025-11-22 08:21:32.305324624 +0000 UTC m=+1113.909663229" Nov 22 08:21:32 crc kubenswrapper[4735]: I1122 08:21:32.375642 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" podStartSLOduration=8.492837444 podStartE2EDuration="56.375620757s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.731611933 +0000 UTC m=+1060.335950538" lastFinishedPulling="2025-11-22 08:21:26.614395236 +0000 UTC m=+1108.218733851" observedRunningTime="2025-11-22 08:21:32.363698436 +0000 UTC m=+1113.968037041" watchObservedRunningTime="2025-11-22 08:21:32.375620757 +0000 UTC m=+1113.979959372" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.230223 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" event={"ID":"2038726f-ef9a-47d2-837e-1b81f72c2d39","Type":"ContainerStarted","Data":"46f2f357ca452aa12897b6aa25a34428cbebc7fcfc337521d28feadcba91ca59"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.230691 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.231787 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" event={"ID":"fc7fded7-7eee-4227-8569-85d07ea4ecd6","Type":"ContainerStarted","Data":"d2b5aa5580d53e5a5be8f7515a4ec8a7a32c650c16add6c25fd8a4477c6eeb31"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.232088 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.233924 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" event={"ID":"2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb","Type":"ContainerStarted","Data":"9f36381e6ba74eb3da1c749ad6f2152ef752a877d053f08dee34e57d46097107"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.234039 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.235784 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" event={"ID":"1c5cb76f-8964-4245-86e5-56b619e06d1b","Type":"ContainerStarted","Data":"2b44951943b3f6c3b64036dfc9d40f08823f047d7f040f019e91987349a56633"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.237519 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" event={"ID":"61871ccb-bdb6-4222-88d9-d34c22a5797c","Type":"ContainerStarted","Data":"bad46aaa8346d4e5c31722710923612edda1accb5c7c17b94472dc74249ee581"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.237709 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.239085 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" event={"ID":"e204b082-d231-48c6-af16-f4ee9d2d7f9c","Type":"ContainerStarted","Data":"c65fd05e36974147428ebfeae9baab5c5c483bff47262bb154b2317ae269f3b7"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.239219 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.240868 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" event={"ID":"2171ac01-b987-4062-9521-88c800930330","Type":"ContainerStarted","Data":"d9e5d9c3ffda2c695c75782097cf0b5e80944c4d8b818408d460de4ac542efce"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.242320 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" event={"ID":"fb88fa45-ef38-49bc-b842-157bef4cd0e5","Type":"ContainerStarted","Data":"3829a9d58cd5cc513826be3b0b9bd684a73dc29ecf08067acc91b4ed608114f6"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.242417 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.244186 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" event={"ID":"10f9fa68-fe17-4af9-8e47-fd534c8a4446","Type":"ContainerStarted","Data":"e2120c5761a31fa34fc9f21b1ea86b676386013266d0df2c4bbd2dac16973931"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.244403 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.246029 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" event={"ID":"c3ac6a85-2f1e-4906-ba7d-c9415f523fa8","Type":"ContainerStarted","Data":"be95ff3a412ffd4deda87e01dcb99c699a8a06c6e608cb53ded2136f31ab968f"} Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.246818 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.248847 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-qs98r" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.249688 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.256655 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-76b989774d-x7gnj" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.268838 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" podStartSLOduration=3.7429856299999997 podStartE2EDuration="58.268819394s" podCreationTimestamp="2025-11-22 08:20:35 +0000 UTC" firstStartedPulling="2025-11-22 08:20:37.521191113 +0000 UTC m=+1059.125529718" lastFinishedPulling="2025-11-22 08:21:32.047024867 +0000 UTC m=+1113.651363482" observedRunningTime="2025-11-22 08:21:33.256904882 +0000 UTC m=+1114.861243487" watchObservedRunningTime="2025-11-22 08:21:33.268819394 +0000 UTC m=+1114.873157999" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.285182 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" podStartSLOduration=3.67716655 podStartE2EDuration="57.285166058s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.444394748 +0000 UTC m=+1060.048733353" lastFinishedPulling="2025-11-22 08:21:32.052394256 +0000 UTC m=+1113.656732861" observedRunningTime="2025-11-22 08:21:33.280237781 +0000 UTC m=+1114.884576386" watchObservedRunningTime="2025-11-22 08:21:33.285166058 +0000 UTC m=+1114.889504663" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.332893 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" podStartSLOduration=3.7114431960000003 podStartE2EDuration="57.332877613s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.414724578 +0000 UTC m=+1060.019063183" lastFinishedPulling="2025-11-22 08:21:32.036158995 +0000 UTC m=+1113.640497600" observedRunningTime="2025-11-22 08:21:33.328915393 +0000 UTC m=+1114.933253998" watchObservedRunningTime="2025-11-22 08:21:33.332877613 +0000 UTC m=+1114.937216218" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.391925 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" podStartSLOduration=3.351469486 podStartE2EDuration="57.391900543s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.000736997 +0000 UTC m=+1059.605075612" lastFinishedPulling="2025-11-22 08:21:32.041168064 +0000 UTC m=+1113.645506669" observedRunningTime="2025-11-22 08:21:33.386348979 +0000 UTC m=+1114.990687584" watchObservedRunningTime="2025-11-22 08:21:33.391900543 +0000 UTC m=+1114.996239148" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.427602 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" podStartSLOduration=3.356860527 podStartE2EDuration="57.427586354s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:37.976424514 +0000 UTC m=+1059.580763119" lastFinishedPulling="2025-11-22 08:21:32.047150331 +0000 UTC m=+1113.651488946" observedRunningTime="2025-11-22 08:21:33.426243557 +0000 UTC m=+1115.030582172" watchObservedRunningTime="2025-11-22 08:21:33.427586354 +0000 UTC m=+1115.031924959" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.450232 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" podStartSLOduration=2.871030347 podStartE2EDuration="57.450207133s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:37.678525745 +0000 UTC m=+1059.282864350" lastFinishedPulling="2025-11-22 08:21:32.257702531 +0000 UTC m=+1113.862041136" observedRunningTime="2025-11-22 08:21:33.442094288 +0000 UTC m=+1115.046432903" watchObservedRunningTime="2025-11-22 08:21:33.450207133 +0000 UTC m=+1115.054545748" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.489011 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" podStartSLOduration=3.413573344 podStartE2EDuration="57.48899478s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:37.976920788 +0000 UTC m=+1059.581259393" lastFinishedPulling="2025-11-22 08:21:32.052342224 +0000 UTC m=+1113.656680829" observedRunningTime="2025-11-22 08:21:33.485793412 +0000 UTC m=+1115.090132017" watchObservedRunningTime="2025-11-22 08:21:33.48899478 +0000 UTC m=+1115.093333385" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.512857 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" podStartSLOduration=29.49023437 podStartE2EDuration="57.512839233s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:21:04.291418622 +0000 UTC m=+1085.895757267" lastFinishedPulling="2025-11-22 08:21:32.314023525 +0000 UTC m=+1113.918362130" observedRunningTime="2025-11-22 08:21:33.511634449 +0000 UTC m=+1115.115973054" watchObservedRunningTime="2025-11-22 08:21:33.512839233 +0000 UTC m=+1115.117177838" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.534159 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" podStartSLOduration=3.838093131 podStartE2EDuration="57.534131964s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.352571948 +0000 UTC m=+1059.956910543" lastFinishedPulling="2025-11-22 08:21:32.048610771 +0000 UTC m=+1113.652949376" observedRunningTime="2025-11-22 08:21:33.525871215 +0000 UTC m=+1115.130209820" watchObservedRunningTime="2025-11-22 08:21:33.534131964 +0000 UTC m=+1115.138470569" Nov 22 08:21:33 crc kubenswrapper[4735]: I1122 08:21:33.545075 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" podStartSLOduration=4.242120949 podStartE2EDuration="57.545057858s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.749424856 +0000 UTC m=+1060.353763471" lastFinishedPulling="2025-11-22 08:21:32.052361775 +0000 UTC m=+1113.656700380" observedRunningTime="2025-11-22 08:21:33.538399553 +0000 UTC m=+1115.142738158" watchObservedRunningTime="2025-11-22 08:21:33.545057858 +0000 UTC m=+1115.149396463" Nov 22 08:21:34 crc kubenswrapper[4735]: I1122 08:21:34.256955 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" Nov 22 08:21:34 crc kubenswrapper[4735]: I1122 08:21:34.257347 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" Nov 22 08:21:35 crc kubenswrapper[4735]: E1122 08:21:35.266340 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" podUID="9124e032-f3c0-46fa-bd17-26ed40bfd57e" Nov 22 08:21:37 crc kubenswrapper[4735]: I1122 08:21:37.071805 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7875d8bb94-tqj8q" Nov 22 08:21:37 crc kubenswrapper[4735]: I1122 08:21:37.108679 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-vndrw" Nov 22 08:21:37 crc kubenswrapper[4735]: I1122 08:21:37.128029 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" Nov 22 08:21:37 crc kubenswrapper[4735]: I1122 08:21:37.136986 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-x7w9w" Nov 22 08:21:37 crc kubenswrapper[4735]: I1122 08:21:37.307240 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-dcgnz" Nov 22 08:21:41 crc kubenswrapper[4735]: I1122 08:21:41.317353 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" event={"ID":"11f148e7-7f11-42cb-ae64-e5714bf56295","Type":"ContainerStarted","Data":"eef656013f60a0486bedb62ff1d32408318c86a04cdf7ec6c3b57b5fccd6b8c7"} Nov 22 08:21:45 crc kubenswrapper[4735]: E1122 08:21:45.267301 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" podUID="5765b990-2731-42c7-bb14-7b149e91555f" Nov 22 08:21:45 crc kubenswrapper[4735]: I1122 08:21:45.321688 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-285tt" podStartSLOduration=30.911645392 podStartE2EDuration="1m9.32166804s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:21:02.356954194 +0000 UTC m=+1083.961292799" lastFinishedPulling="2025-11-22 08:21:40.766976842 +0000 UTC m=+1122.371315447" observedRunningTime="2025-11-22 08:21:41.334176621 +0000 UTC m=+1122.938515236" watchObservedRunningTime="2025-11-22 08:21:45.32166804 +0000 UTC m=+1126.926006665" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.131582 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.132048 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.132129 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.133408 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33c0e7ebc11dc73bdfa42a0d62c634dcd145b69c651d39f5221e590919439f72"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.133552 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://33c0e7ebc11dc73bdfa42a0d62c634dcd145b69c651d39f5221e590919439f72" gracePeriod=600 Nov 22 08:21:46 crc kubenswrapper[4735]: E1122 08:21:46.287400 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" podUID="150b26f8-b5cd-4828-96cc-6e37000c24e7" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.362655 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.378980 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" event={"ID":"990c8bbe-2866-4568-8e30-1d9f016f0736","Type":"ContainerStarted","Data":"caa61b448d8623a53f9257246b7fe046111c2591dc31af2655d55c0a5c30bc61"} Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.379332 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.399126 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="33c0e7ebc11dc73bdfa42a0d62c634dcd145b69c651d39f5221e590919439f72" exitCode=0 Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.399184 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"33c0e7ebc11dc73bdfa42a0d62c634dcd145b69c651d39f5221e590919439f72"} Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.399251 4735 scope.go:117] "RemoveContainer" containerID="8ac3677e0ebb55329e2a912a3a81033db8ba05ef418621a9d574aafba4dff19d" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.420836 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.424672 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" podStartSLOduration=28.932262466 podStartE2EDuration="1m10.424649176s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:21:04.293039606 +0000 UTC m=+1085.897378251" lastFinishedPulling="2025-11-22 08:21:45.785426356 +0000 UTC m=+1127.389764961" observedRunningTime="2025-11-22 08:21:46.418739422 +0000 UTC m=+1128.023078047" watchObservedRunningTime="2025-11-22 08:21:46.424649176 +0000 UTC m=+1128.028987781" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.446867 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.497423 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.567827 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.571921 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" Nov 22 08:21:46 crc kubenswrapper[4735]: I1122 08:21:46.873847 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2kf8g" Nov 22 08:21:47 crc kubenswrapper[4735]: I1122 08:21:47.408793 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" event={"ID":"fce6a4d4-f6fa-44bf-8528-7de867925bbb","Type":"ContainerStarted","Data":"d81e8021c482422a6f57a7b45762231f1205554bc61bf64af412d58bc7e0d086"} Nov 22 08:21:47 crc kubenswrapper[4735]: I1122 08:21:47.409271 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" Nov 22 08:21:47 crc kubenswrapper[4735]: I1122 08:21:47.411743 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"4afcbebd71a886d2b368e2d15dd469e8f84d1fd0d397678cdfa5ebd38afdfac1"} Nov 22 08:21:47 crc kubenswrapper[4735]: I1122 08:21:47.414497 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" event={"ID":"ee3c788a-51f9-4808-996b-e854ba26a776","Type":"ContainerStarted","Data":"9f1d651f3f56e597c1430b918b0b19222834769d37fc2ec583b148447f529a97"} Nov 22 08:21:47 crc kubenswrapper[4735]: I1122 08:21:47.414643 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" Nov 22 08:21:47 crc kubenswrapper[4735]: I1122 08:21:47.427623 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" podStartSLOduration=3.342550798 podStartE2EDuration="1m11.427609542s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.732029705 +0000 UTC m=+1060.336368310" lastFinishedPulling="2025-11-22 08:21:46.817088439 +0000 UTC m=+1128.421427054" observedRunningTime="2025-11-22 08:21:47.424439224 +0000 UTC m=+1129.028777829" watchObservedRunningTime="2025-11-22 08:21:47.427609542 +0000 UTC m=+1129.031948147" Nov 22 08:21:47 crc kubenswrapper[4735]: I1122 08:21:47.464764 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" podStartSLOduration=3.842575688 podStartE2EDuration="1m11.464747294s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.744731015 +0000 UTC m=+1060.349069620" lastFinishedPulling="2025-11-22 08:21:46.366902621 +0000 UTC m=+1127.971241226" observedRunningTime="2025-11-22 08:21:47.458696026 +0000 UTC m=+1129.063034631" watchObservedRunningTime="2025-11-22 08:21:47.464747294 +0000 UTC m=+1129.069085899" Nov 22 08:21:48 crc kubenswrapper[4735]: I1122 08:21:48.423208 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" event={"ID":"3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c","Type":"ContainerStarted","Data":"b58753515763e9650186a1df4f80f44de7c3cd5a8fcc24ce072d2f968aaa0eda"} Nov 22 08:21:48 crc kubenswrapper[4735]: I1122 08:21:48.424680 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" Nov 22 08:21:48 crc kubenswrapper[4735]: I1122 08:21:48.426449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" event={"ID":"2c87c2e3-b68e-4f75-a7cf-2c752e7286e8","Type":"ContainerStarted","Data":"9f5b798326f30cae0cef308b1d613040ecf9fd9e632ad4cc915aec5d576aa055"} Nov 22 08:21:48 crc kubenswrapper[4735]: I1122 08:21:48.426849 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" Nov 22 08:21:48 crc kubenswrapper[4735]: I1122 08:21:48.439282 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" podStartSLOduration=3.107395512 podStartE2EDuration="1m12.43926891s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.414315546 +0000 UTC m=+1060.018654161" lastFinishedPulling="2025-11-22 08:21:47.746188954 +0000 UTC m=+1129.350527559" observedRunningTime="2025-11-22 08:21:48.436618036 +0000 UTC m=+1130.040956661" watchObservedRunningTime="2025-11-22 08:21:48.43926891 +0000 UTC m=+1130.043607515" Nov 22 08:21:48 crc kubenswrapper[4735]: I1122 08:21:48.467067 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" podStartSLOduration=4.062912828 podStartE2EDuration="1m13.467046652s" podCreationTimestamp="2025-11-22 08:20:35 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.407811226 +0000 UTC m=+1060.012149831" lastFinishedPulling="2025-11-22 08:21:47.81194505 +0000 UTC m=+1129.416283655" observedRunningTime="2025-11-22 08:21:48.460115909 +0000 UTC m=+1130.064454554" watchObservedRunningTime="2025-11-22 08:21:48.467046652 +0000 UTC m=+1130.071385257" Nov 22 08:21:49 crc kubenswrapper[4735]: I1122 08:21:49.437559 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" event={"ID":"9124e032-f3c0-46fa-bd17-26ed40bfd57e","Type":"ContainerStarted","Data":"720595dda19bf388fd3cf948cf7a368d0a64caf84ad85114e8d647ba93c0145c"} Nov 22 08:21:49 crc kubenswrapper[4735]: I1122 08:21:49.438578 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" Nov 22 08:21:49 crc kubenswrapper[4735]: I1122 08:21:49.456011 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" podStartSLOduration=3.5246102820000003 podStartE2EDuration="1m13.455990629s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.770816288 +0000 UTC m=+1060.375154893" lastFinishedPulling="2025-11-22 08:21:48.702196625 +0000 UTC m=+1130.306535240" observedRunningTime="2025-11-22 08:21:49.450887567 +0000 UTC m=+1131.055226182" watchObservedRunningTime="2025-11-22 08:21:49.455990629 +0000 UTC m=+1131.060329234" Nov 22 08:21:56 crc kubenswrapper[4735]: I1122 08:21:56.610225 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" Nov 22 08:21:56 crc kubenswrapper[4735]: I1122 08:21:56.775199 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" Nov 22 08:21:56 crc kubenswrapper[4735]: I1122 08:21:56.790060 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" Nov 22 08:21:57 crc kubenswrapper[4735]: I1122 08:21:57.228022 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7d76d46c6f-shjbx" Nov 22 08:21:57 crc kubenswrapper[4735]: I1122 08:21:57.405958 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cn8lg" Nov 22 08:21:57 crc kubenswrapper[4735]: I1122 08:21:57.557865 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" event={"ID":"5765b990-2731-42c7-bb14-7b149e91555f","Type":"ContainerStarted","Data":"ce66a9c7fdad3e1f5c6445dcd6efdbbc34a286be58a0e16a29f65c7d01ffca70"} Nov 22 08:21:57 crc kubenswrapper[4735]: I1122 08:21:57.558120 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" Nov 22 08:21:57 crc kubenswrapper[4735]: I1122 08:21:57.584116 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" podStartSLOduration=3.312338532 podStartE2EDuration="1m22.58409673s" podCreationTimestamp="2025-11-22 08:20:35 +0000 UTC" firstStartedPulling="2025-11-22 08:20:37.447404891 +0000 UTC m=+1059.051743496" lastFinishedPulling="2025-11-22 08:21:56.719163099 +0000 UTC m=+1138.323501694" observedRunningTime="2025-11-22 08:21:57.578968288 +0000 UTC m=+1139.183306893" watchObservedRunningTime="2025-11-22 08:21:57.58409673 +0000 UTC m=+1139.188435325" Nov 22 08:21:57 crc kubenswrapper[4735]: I1122 08:21:57.634975 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc" Nov 22 08:21:59 crc kubenswrapper[4735]: I1122 08:21:59.581180 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" event={"ID":"150b26f8-b5cd-4828-96cc-6e37000c24e7","Type":"ContainerStarted","Data":"ddead189a96cff5eb6367b3029c7e9bc9b42927b97dbb20043eb6c2a3d9070e1"} Nov 22 08:21:59 crc kubenswrapper[4735]: I1122 08:21:59.581668 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" Nov 22 08:21:59 crc kubenswrapper[4735]: I1122 08:21:59.602367 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" podStartSLOduration=3.274235199 podStartE2EDuration="1m23.602351676s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="2025-11-22 08:20:38.414640056 +0000 UTC m=+1060.018978661" lastFinishedPulling="2025-11-22 08:21:58.742756533 +0000 UTC m=+1140.347095138" observedRunningTime="2025-11-22 08:21:59.598661233 +0000 UTC m=+1141.202999838" watchObservedRunningTime="2025-11-22 08:21:59.602351676 +0000 UTC m=+1141.206690281" Nov 22 08:22:06 crc kubenswrapper[4735]: I1122 08:22:06.322251 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" Nov 22 08:22:06 crc kubenswrapper[4735]: I1122 08:22:06.822074 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-b47vs" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.168630 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9lbg4"] Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.170815 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.174333 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.174657 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.174793 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.174914 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-kfw7v" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.193270 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9lbg4"] Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.255469 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55wh7"] Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.256778 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.258805 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.274534 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55wh7"] Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.287602 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29dql\" (UniqueName: \"kubernetes.io/projected/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-kube-api-access-29dql\") pod \"dnsmasq-dns-675f4bcbfc-9lbg4\" (UID: \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.287667 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-config\") pod \"dnsmasq-dns-675f4bcbfc-9lbg4\" (UID: \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.389639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-config\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.389708 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.389744 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz29v\" (UniqueName: \"kubernetes.io/projected/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-kube-api-access-gz29v\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.390018 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29dql\" (UniqueName: \"kubernetes.io/projected/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-kube-api-access-29dql\") pod \"dnsmasq-dns-675f4bcbfc-9lbg4\" (UID: \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.390066 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-config\") pod \"dnsmasq-dns-675f4bcbfc-9lbg4\" (UID: \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.390769 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-config\") pod \"dnsmasq-dns-675f4bcbfc-9lbg4\" (UID: \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.410655 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29dql\" (UniqueName: \"kubernetes.io/projected/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-kube-api-access-29dql\") pod \"dnsmasq-dns-675f4bcbfc-9lbg4\" (UID: \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.491117 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-config\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.491182 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.491225 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz29v\" (UniqueName: \"kubernetes.io/projected/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-kube-api-access-gz29v\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.491995 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.492302 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-config\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.493453 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.516815 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz29v\" (UniqueName: \"kubernetes.io/projected/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-kube-api-access-gz29v\") pod \"dnsmasq-dns-78dd6ddcc-55wh7\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.576843 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:23 crc kubenswrapper[4735]: I1122 08:22:23.981733 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9lbg4"] Nov 22 08:22:24 crc kubenswrapper[4735]: I1122 08:22:24.080321 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55wh7"] Nov 22 08:22:24 crc kubenswrapper[4735]: W1122 08:22:24.082678 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4697ffd7_6b0c_484b_acda_110f0ffb5a5d.slice/crio-4dd6ead2feab6adc705bdb213ae56cda7a4884c15e644036e739ff233204179b WatchSource:0}: Error finding container 4dd6ead2feab6adc705bdb213ae56cda7a4884c15e644036e739ff233204179b: Status 404 returned error can't find the container with id 4dd6ead2feab6adc705bdb213ae56cda7a4884c15e644036e739ff233204179b Nov 22 08:22:24 crc kubenswrapper[4735]: I1122 08:22:24.836083 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" event={"ID":"8f9d9209-ed3c-49fa-83cb-bdaa27a65487","Type":"ContainerStarted","Data":"9b4a60522141c197454ce890a0ae2ffdf7519b5013448dde734f04af44229f64"} Nov 22 08:22:24 crc kubenswrapper[4735]: I1122 08:22:24.838427 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" event={"ID":"4697ffd7-6b0c-484b-acda-110f0ffb5a5d","Type":"ContainerStarted","Data":"4dd6ead2feab6adc705bdb213ae56cda7a4884c15e644036e739ff233204179b"} Nov 22 08:22:25 crc kubenswrapper[4735]: I1122 08:22:25.905289 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9lbg4"] Nov 22 08:22:25 crc kubenswrapper[4735]: I1122 08:22:25.927019 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ql9fl"] Nov 22 08:22:25 crc kubenswrapper[4735]: I1122 08:22:25.928880 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:25 crc kubenswrapper[4735]: I1122 08:22:25.946366 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-config\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:25 crc kubenswrapper[4735]: I1122 08:22:25.946432 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ml2v\" (UniqueName: \"kubernetes.io/projected/4abed129-55b7-4d35-99e4-d8817c7dc39e-kube-api-access-9ml2v\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:25 crc kubenswrapper[4735]: I1122 08:22:25.946589 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:25 crc kubenswrapper[4735]: I1122 08:22:25.947201 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ql9fl"] Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.050404 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.050557 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-config\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.050604 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ml2v\" (UniqueName: \"kubernetes.io/projected/4abed129-55b7-4d35-99e4-d8817c7dc39e-kube-api-access-9ml2v\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.052001 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-config\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.052002 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.100366 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ml2v\" (UniqueName: \"kubernetes.io/projected/4abed129-55b7-4d35-99e4-d8817c7dc39e-kube-api-access-9ml2v\") pod \"dnsmasq-dns-666b6646f7-ql9fl\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.221170 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55wh7"] Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.249956 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8rp4w"] Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.251955 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.257334 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.259536 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8rp4w"] Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.365806 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkl9p\" (UniqueName: \"kubernetes.io/projected/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-kube-api-access-kkl9p\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.366069 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.366123 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.468881 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.469636 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.469888 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkl9p\" (UniqueName: \"kubernetes.io/projected/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-kube-api-access-kkl9p\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.469928 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.470515 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.502523 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkl9p\" (UniqueName: \"kubernetes.io/projected/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-kube-api-access-kkl9p\") pod \"dnsmasq-dns-57d769cc4f-8rp4w\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.677065 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:22:26 crc kubenswrapper[4735]: I1122 08:22:26.877338 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ql9fl"] Nov 22 08:22:26 crc kubenswrapper[4735]: W1122 08:22:26.885899 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4abed129_55b7_4d35_99e4_d8817c7dc39e.slice/crio-0900d3d2c8f811738df1aa2faecaad36597ec988dcf983141f8553d496b156e7 WatchSource:0}: Error finding container 0900d3d2c8f811738df1aa2faecaad36597ec988dcf983141f8553d496b156e7: Status 404 returned error can't find the container with id 0900d3d2c8f811738df1aa2faecaad36597ec988dcf983141f8553d496b156e7 Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.083100 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.084567 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.087816 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.088261 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.088386 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.088596 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.088756 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.088867 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.090181 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mk25c" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.104579 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.185876 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.185947 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-config-data\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.185981 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.186008 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.186030 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.186075 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgb2t\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-kube-api-access-lgb2t\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.186096 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.186120 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.186146 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.186199 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.186220 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.290681 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.291896 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292034 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292069 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292274 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292387 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-config-data\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292490 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292549 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292576 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292651 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgb2t\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-kube-api-access-lgb2t\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.292678 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.293782 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.294428 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.294594 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-config-data\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.294632 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.294673 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.297231 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.297951 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.299767 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.300053 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8rp4w"] Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.305089 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.315914 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.318223 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgb2t\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-kube-api-access-lgb2t\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.334419 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.399727 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.401298 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.406017 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.406299 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.406444 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.406777 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.407447 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.408269 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.408527 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-j67jk" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.424758 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.451979 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.606840 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ee793a0-c437-47a3-b05d-369a1730d6b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.606882 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.606923 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.607007 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.607052 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.607084 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.607109 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f6c2\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-kube-api-access-4f6c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.607137 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.607167 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.607200 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ee793a0-c437-47a3-b05d-369a1730d6b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.607225 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.708828 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.709195 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.709241 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.709268 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f6c2\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-kube-api-access-4f6c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.709294 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.709328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.709409 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ee793a0-c437-47a3-b05d-369a1730d6b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.709477 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.709990 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.710048 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ee793a0-c437-47a3-b05d-369a1730d6b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.710136 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.711674 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.712037 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.713280 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.713414 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.715090 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.715994 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.716564 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.717357 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ee793a0-c437-47a3-b05d-369a1730d6b1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.730428 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.730966 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ee793a0-c437-47a3-b05d-369a1730d6b1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.741581 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f6c2\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-kube-api-access-4f6c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.765389 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.871321 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" event={"ID":"4abed129-55b7-4d35-99e4-d8817c7dc39e","Type":"ContainerStarted","Data":"0900d3d2c8f811738df1aa2faecaad36597ec988dcf983141f8553d496b156e7"} Nov 22 08:22:27 crc kubenswrapper[4735]: I1122 08:22:27.875453 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" event={"ID":"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb","Type":"ContainerStarted","Data":"3471102029aa054e8554ef597d7a9718139712ae51aa6c22ca4174aad9c9db08"} Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.038864 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.057796 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.723611 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:22:28 crc kubenswrapper[4735]: W1122 08:22:28.760096 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ee793a0_c437_47a3_b05d_369a1730d6b1.slice/crio-f01069603ce7dd2091a4a771d1c6acc21d979dd103325c8234adabcb49a87dc8 WatchSource:0}: Error finding container f01069603ce7dd2091a4a771d1c6acc21d979dd103325c8234adabcb49a87dc8: Status 404 returned error can't find the container with id f01069603ce7dd2091a4a771d1c6acc21d979dd103325c8234adabcb49a87dc8 Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.770478 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.772388 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.776759 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.777088 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.777240 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-pb48k" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.777380 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.784659 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.796677 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.892100 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5ee793a0-c437-47a3-b05d-369a1730d6b1","Type":"ContainerStarted","Data":"f01069603ce7dd2091a4a771d1c6acc21d979dd103325c8234adabcb49a87dc8"} Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.898895 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10","Type":"ContainerStarted","Data":"818516c632fa956d6867999897cd1febf5bd027c197923a272842e62e8aa06ae"} Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.938530 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.938598 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28hfh\" (UniqueName: \"kubernetes.io/projected/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-kube-api-access-28hfh\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.938666 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.938722 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.938827 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-config-data-default\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.938976 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.939029 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:28 crc kubenswrapper[4735]: I1122 08:22:28.939071 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-kolla-config\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.044703 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.044765 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.044799 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-config-data-default\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.044850 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.044870 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.044891 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-kolla-config\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.044935 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.044957 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28hfh\" (UniqueName: \"kubernetes.io/projected/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-kube-api-access-28hfh\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.046388 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.047965 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.048718 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-config-data-default\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.049169 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-kolla-config\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.053203 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.063281 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.100420 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.161766 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.163920 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28hfh\" (UniqueName: \"kubernetes.io/projected/5eaebfb3-8113-41d6-9a3d-77c8194ad4d1-kube-api-access-28hfh\") pod \"openstack-galera-0\" (UID: \"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1\") " pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.399451 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 22 08:22:29 crc kubenswrapper[4735]: I1122 08:22:29.983026 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.305995 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.311351 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.314853 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-pqfqz" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.317440 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.317732 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.317864 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.333258 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.430393 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.430447 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.430513 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c715895-af70-4e8f-96c4-c7064e45c2c9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.430703 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.430956 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0c715895-af70-4e8f-96c4-c7064e45c2c9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.431151 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.431349 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c715895-af70-4e8f-96c4-c7064e45c2c9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.431441 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbt7w\" (UniqueName: \"kubernetes.io/projected/0c715895-af70-4e8f-96c4-c7064e45c2c9-kube-api-access-qbt7w\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.509853 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.514838 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.518340 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-w4cqd" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.518603 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.518902 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.533782 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c715895-af70-4e8f-96c4-c7064e45c2c9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.533830 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbt7w\" (UniqueName: \"kubernetes.io/projected/0c715895-af70-4e8f-96c4-c7064e45c2c9-kube-api-access-qbt7w\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.533873 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.533893 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.533915 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c715895-af70-4e8f-96c4-c7064e45c2c9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.533958 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.534007 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0c715895-af70-4e8f-96c4-c7064e45c2c9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.534050 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.535050 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.535051 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0c715895-af70-4e8f-96c4-c7064e45c2c9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.535647 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.536181 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.541179 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c715895-af70-4e8f-96c4-c7064e45c2c9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.542191 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c715895-af70-4e8f-96c4-c7064e45c2c9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.546940 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.547031 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0c715895-af70-4e8f-96c4-c7064e45c2c9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.577971 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbt7w\" (UniqueName: \"kubernetes.io/projected/0c715895-af70-4e8f-96c4-c7064e45c2c9-kube-api-access-qbt7w\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.609441 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0c715895-af70-4e8f-96c4-c7064e45c2c9\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.642712 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvw89\" (UniqueName: \"kubernetes.io/projected/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-kube-api-access-mvw89\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.642773 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-config-data\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.642854 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-kolla-config\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.642871 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.642915 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.669908 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.749822 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.749919 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvw89\" (UniqueName: \"kubernetes.io/projected/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-kube-api-access-mvw89\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.749953 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-config-data\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.750032 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-kolla-config\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.750050 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.752521 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-kolla-config\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.752701 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-config-data\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.753560 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.756673 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.773548 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvw89\" (UniqueName: \"kubernetes.io/projected/b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef-kube-api-access-mvw89\") pod \"memcached-0\" (UID: \"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef\") " pod="openstack/memcached-0" Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.945850 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1","Type":"ContainerStarted","Data":"883f526037266f8ddbf006a844c62eeaa07961bb0db8eff80fabcbcf376d249f"} Nov 22 08:22:30 crc kubenswrapper[4735]: I1122 08:22:30.964342 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 22 08:22:31 crc kubenswrapper[4735]: I1122 08:22:31.256620 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 08:22:31 crc kubenswrapper[4735]: I1122 08:22:31.561957 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 22 08:22:32 crc kubenswrapper[4735]: I1122 08:22:32.547025 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:22:32 crc kubenswrapper[4735]: I1122 08:22:32.548731 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:22:32 crc kubenswrapper[4735]: I1122 08:22:32.558419 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-bn5m6" Nov 22 08:22:32 crc kubenswrapper[4735]: I1122 08:22:32.571307 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:22:32 crc kubenswrapper[4735]: I1122 08:22:32.700535 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgmkz\" (UniqueName: \"kubernetes.io/projected/4c7911f8-389a-43fe-a011-dcc65ccbd7f2-kube-api-access-dgmkz\") pod \"kube-state-metrics-0\" (UID: \"4c7911f8-389a-43fe-a011-dcc65ccbd7f2\") " pod="openstack/kube-state-metrics-0" Nov 22 08:22:32 crc kubenswrapper[4735]: I1122 08:22:32.806808 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgmkz\" (UniqueName: \"kubernetes.io/projected/4c7911f8-389a-43fe-a011-dcc65ccbd7f2-kube-api-access-dgmkz\") pod \"kube-state-metrics-0\" (UID: \"4c7911f8-389a-43fe-a011-dcc65ccbd7f2\") " pod="openstack/kube-state-metrics-0" Nov 22 08:22:32 crc kubenswrapper[4735]: I1122 08:22:32.832217 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgmkz\" (UniqueName: \"kubernetes.io/projected/4c7911f8-389a-43fe-a011-dcc65ccbd7f2-kube-api-access-dgmkz\") pod \"kube-state-metrics-0\" (UID: \"4c7911f8-389a-43fe-a011-dcc65ccbd7f2\") " pod="openstack/kube-state-metrics-0" Nov 22 08:22:32 crc kubenswrapper[4735]: I1122 08:22:32.894034 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.350663 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c"] Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.351975 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.360695 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-nfmwb" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.360996 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.365375 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c"] Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.523316 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36da6203-0695-45d5-b361-01267bd61a1d-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-2dl4c\" (UID: \"36da6203-0695-45d5-b361-01267bd61a1d\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.523435 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phzm8\" (UniqueName: \"kubernetes.io/projected/36da6203-0695-45d5-b361-01267bd61a1d-kube-api-access-phzm8\") pod \"observability-ui-dashboards-7d5fb4cbfb-2dl4c\" (UID: \"36da6203-0695-45d5-b361-01267bd61a1d\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.625401 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phzm8\" (UniqueName: \"kubernetes.io/projected/36da6203-0695-45d5-b361-01267bd61a1d-kube-api-access-phzm8\") pod \"observability-ui-dashboards-7d5fb4cbfb-2dl4c\" (UID: \"36da6203-0695-45d5-b361-01267bd61a1d\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.625666 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36da6203-0695-45d5-b361-01267bd61a1d-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-2dl4c\" (UID: \"36da6203-0695-45d5-b361-01267bd61a1d\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:33 crc kubenswrapper[4735]: E1122 08:22:33.625791 4735 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Nov 22 08:22:33 crc kubenswrapper[4735]: E1122 08:22:33.625839 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36da6203-0695-45d5-b361-01267bd61a1d-serving-cert podName:36da6203-0695-45d5-b361-01267bd61a1d nodeName:}" failed. No retries permitted until 2025-11-22 08:22:34.125823694 +0000 UTC m=+1175.730162299 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/36da6203-0695-45d5-b361-01267bd61a1d-serving-cert") pod "observability-ui-dashboards-7d5fb4cbfb-2dl4c" (UID: "36da6203-0695-45d5-b361-01267bd61a1d") : secret "observability-ui-dashboards" not found Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.656438 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5df59bb7db-gjtgp"] Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.665152 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phzm8\" (UniqueName: \"kubernetes.io/projected/36da6203-0695-45d5-b361-01267bd61a1d-kube-api-access-phzm8\") pod \"observability-ui-dashboards-7d5fb4cbfb-2dl4c\" (UID: \"36da6203-0695-45d5-b361-01267bd61a1d\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.674493 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.682937 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5df59bb7db-gjtgp"] Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.828964 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-oauth-config\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.829056 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4twn7\" (UniqueName: \"kubernetes.io/projected/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-kube-api-access-4twn7\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.829083 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-oauth-serving-cert\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.829102 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-config\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.829126 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-trusted-ca-bundle\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.829178 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-service-ca\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.829244 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-serving-cert\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.930868 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-service-ca\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.931005 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-serving-cert\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.931060 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-oauth-config\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.931166 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4twn7\" (UniqueName: \"kubernetes.io/projected/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-kube-api-access-4twn7\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.931204 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-oauth-serving-cert\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.931239 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-config\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.931268 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-trusted-ca-bundle\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.932194 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-trusted-ca-bundle\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.932290 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-oauth-serving-cert\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.931761 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-service-ca\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.932930 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-config\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.938050 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-oauth-config\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.938628 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-console-serving-cert\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:33 crc kubenswrapper[4735]: I1122 08:22:33.950435 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4twn7\" (UniqueName: \"kubernetes.io/projected/0e0587d6-2a11-4bb9-9be8-ac139a2ff912-kube-api-access-4twn7\") pod \"console-5df59bb7db-gjtgp\" (UID: \"0e0587d6-2a11-4bb9-9be8-ac139a2ff912\") " pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.010413 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.134872 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36da6203-0695-45d5-b361-01267bd61a1d-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-2dl4c\" (UID: \"36da6203-0695-45d5-b361-01267bd61a1d\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.154197 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36da6203-0695-45d5-b361-01267bd61a1d-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-2dl4c\" (UID: \"36da6203-0695-45d5-b361-01267bd61a1d\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.187692 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.203284 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.205803 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.206033 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-gd8n8" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.206221 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.206609 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.208443 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.210424 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.216939 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.274955 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.337610 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj4ll\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-kube-api-access-jj4ll\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.337684 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.337740 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/bea30a30-b625-4511-87da-f26e5495a5f9-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.337758 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-config\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.337795 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/bea30a30-b625-4511-87da-f26e5495a5f9-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.337874 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.337941 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.337974 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.439251 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.439522 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.439568 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj4ll\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-kube-api-access-jj4ll\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.439602 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.439638 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/bea30a30-b625-4511-87da-f26e5495a5f9-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.439654 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-config\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.439672 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/bea30a30-b625-4511-87da-f26e5495a5f9-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.439733 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.440708 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/bea30a30-b625-4511-87da-f26e5495a5f9-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.443520 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.443519 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-config\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.443875 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.443909 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/504ad61e9b31def24aaaccbb19a5701f8cf89a1c3574104486b5c4b008bcacd8/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.446767 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.461782 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/bea30a30-b625-4511-87da-f26e5495a5f9-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.466276 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj4ll\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-kube-api-access-jj4ll\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.466765 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.487034 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"prometheus-metric-storage-0\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:34 crc kubenswrapper[4735]: I1122 08:22:34.532707 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.860594 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z5csg"] Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.861949 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.869790 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.873034 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.873224 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-jmlq8" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.880047 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7mhwj"] Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.882387 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.910641 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z5csg"] Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.922823 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7mhwj"] Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974718 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-etc-ovs\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974772 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fee31945-9fca-4693-9218-97bb4513cee3-scripts\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974799 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-log\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974832 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-lib\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974850 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-run-ovn\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d99e6f50-a748-4ae8-b1b0-9df901b37201-ovn-controller-tls-certs\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974918 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkssj\" (UniqueName: \"kubernetes.io/projected/fee31945-9fca-4693-9218-97bb4513cee3-kube-api-access-qkssj\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-log-ovn\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.974997 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d99e6f50-a748-4ae8-b1b0-9df901b37201-combined-ca-bundle\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.975057 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj7jd\" (UniqueName: \"kubernetes.io/projected/d99e6f50-a748-4ae8-b1b0-9df901b37201-kube-api-access-cj7jd\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.975076 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d99e6f50-a748-4ae8-b1b0-9df901b37201-scripts\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.975095 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-run\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:35 crc kubenswrapper[4735]: I1122 08:22:35.975122 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-run\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076637 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj7jd\" (UniqueName: \"kubernetes.io/projected/d99e6f50-a748-4ae8-b1b0-9df901b37201-kube-api-access-cj7jd\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076693 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d99e6f50-a748-4ae8-b1b0-9df901b37201-scripts\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076721 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-run\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076771 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-run\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076809 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-etc-ovs\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076832 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fee31945-9fca-4693-9218-97bb4513cee3-scripts\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076863 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-log\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076900 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-lib\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076918 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-run-ovn\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.076961 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d99e6f50-a748-4ae8-b1b0-9df901b37201-ovn-controller-tls-certs\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.077018 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkssj\" (UniqueName: \"kubernetes.io/projected/fee31945-9fca-4693-9218-97bb4513cee3-kube-api-access-qkssj\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.077065 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-log-ovn\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.077081 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d99e6f50-a748-4ae8-b1b0-9df901b37201-combined-ca-bundle\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.077498 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-run\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.079901 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d99e6f50-a748-4ae8-b1b0-9df901b37201-scripts\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.080060 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-lib\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.080111 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-run\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.080210 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-etc-ovs\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.080538 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-run-ovn\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.080645 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fee31945-9fca-4693-9218-97bb4513cee3-var-log\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.080691 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d99e6f50-a748-4ae8-b1b0-9df901b37201-var-log-ovn\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.086437 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d99e6f50-a748-4ae8-b1b0-9df901b37201-combined-ca-bundle\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.086634 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d99e6f50-a748-4ae8-b1b0-9df901b37201-ovn-controller-tls-certs\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.088799 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fee31945-9fca-4693-9218-97bb4513cee3-scripts\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.101037 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj7jd\" (UniqueName: \"kubernetes.io/projected/d99e6f50-a748-4ae8-b1b0-9df901b37201-kube-api-access-cj7jd\") pod \"ovn-controller-z5csg\" (UID: \"d99e6f50-a748-4ae8-b1b0-9df901b37201\") " pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.123955 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkssj\" (UniqueName: \"kubernetes.io/projected/fee31945-9fca-4693-9218-97bb4513cee3-kube-api-access-qkssj\") pod \"ovn-controller-ovs-7mhwj\" (UID: \"fee31945-9fca-4693-9218-97bb4513cee3\") " pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.185760 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.220743 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.761663 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.763720 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.766385 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.766710 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.766856 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.767081 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qrr62" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.767734 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.770446 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.894529 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.894722 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.894885 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-config\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.894941 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.895135 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.895335 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.895588 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kr45\" (UniqueName: \"kubernetes.io/projected/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-kube-api-access-6kr45\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.895673 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.997824 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-config\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.997880 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.997928 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.997981 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.998045 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kr45\" (UniqueName: \"kubernetes.io/projected/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-kube-api-access-6kr45\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.998073 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.998105 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.998140 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.998309 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:36 crc kubenswrapper[4735]: I1122 08:22:36.998901 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-config\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:37 crc kubenswrapper[4735]: I1122 08:22:37.000567 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:37 crc kubenswrapper[4735]: I1122 08:22:37.003863 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:37 crc kubenswrapper[4735]: I1122 08:22:37.004902 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:37 crc kubenswrapper[4735]: I1122 08:22:37.007316 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:37 crc kubenswrapper[4735]: I1122 08:22:37.008104 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:37 crc kubenswrapper[4735]: I1122 08:22:37.026154 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kr45\" (UniqueName: \"kubernetes.io/projected/135fc91d-a8bd-489a-a7ae-3df9db9af6cd-kube-api-access-6kr45\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:37 crc kubenswrapper[4735]: I1122 08:22:37.027393 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"135fc91d-a8bd-489a-a7ae-3df9db9af6cd\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:37 crc kubenswrapper[4735]: I1122 08:22:37.097031 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 22 08:22:38 crc kubenswrapper[4735]: W1122 08:22:38.949772 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c715895_af70_4e8f_96c4_c7064e45c2c9.slice/crio-42ba2e9163ad93553bfb53276e93032e171af2659f40cccc49783c80b9b954ef WatchSource:0}: Error finding container 42ba2e9163ad93553bfb53276e93032e171af2659f40cccc49783c80b9b954ef: Status 404 returned error can't find the container with id 42ba2e9163ad93553bfb53276e93032e171af2659f40cccc49783c80b9b954ef Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.080973 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef","Type":"ContainerStarted","Data":"3d3bb54d3fd88f0871a0d22013a91e607dcbbcdc668e1c299127d17d0b64bfd9"} Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.081907 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0c715895-af70-4e8f-96c4-c7064e45c2c9","Type":"ContainerStarted","Data":"42ba2e9163ad93553bfb53276e93032e171af2659f40cccc49783c80b9b954ef"} Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.593855 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.595973 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.597430 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-k9qlh" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.599501 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.599690 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.601678 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.606284 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.751761 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.751884 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.751949 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16a0218b-2a6b-47f5-832f-2d8da9f950f3-config\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.751965 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7dn2\" (UniqueName: \"kubernetes.io/projected/16a0218b-2a6b-47f5-832f-2d8da9f950f3-kube-api-access-l7dn2\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.752026 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.752043 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16a0218b-2a6b-47f5-832f-2d8da9f950f3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.752059 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/16a0218b-2a6b-47f5-832f-2d8da9f950f3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.752097 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854119 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854211 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854272 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16a0218b-2a6b-47f5-832f-2d8da9f950f3-config\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854295 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7dn2\" (UniqueName: \"kubernetes.io/projected/16a0218b-2a6b-47f5-832f-2d8da9f950f3-kube-api-access-l7dn2\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854355 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854379 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16a0218b-2a6b-47f5-832f-2d8da9f950f3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854401 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/16a0218b-2a6b-47f5-832f-2d8da9f950f3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854408 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.854428 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.855370 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/16a0218b-2a6b-47f5-832f-2d8da9f950f3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.857832 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16a0218b-2a6b-47f5-832f-2d8da9f950f3-config\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.858665 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16a0218b-2a6b-47f5-832f-2d8da9f950f3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.868191 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.868493 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.870196 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a0218b-2a6b-47f5-832f-2d8da9f950f3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.876756 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7dn2\" (UniqueName: \"kubernetes.io/projected/16a0218b-2a6b-47f5-832f-2d8da9f950f3-kube-api-access-l7dn2\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.883100 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"16a0218b-2a6b-47f5-832f-2d8da9f950f3\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:39 crc kubenswrapper[4735]: I1122 08:22:39.929148 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 22 08:22:47 crc kubenswrapper[4735]: E1122 08:22:47.090547 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 22 08:22:47 crc kubenswrapper[4735]: E1122 08:22:47.091782 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4f6c2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(5ee793a0-c437-47a3-b05d-369a1730d6b1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:22:47 crc kubenswrapper[4735]: E1122 08:22:47.093063 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" Nov 22 08:22:47 crc kubenswrapper[4735]: E1122 08:22:47.151131 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.227733 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.228477 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-29dql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-9lbg4_openstack(8f9d9209-ed3c-49fa-83cb-bdaa27a65487): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.229736 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" podUID="8f9d9209-ed3c-49fa-83cb-bdaa27a65487" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.294297 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.294697 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gz29v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-55wh7_openstack(4697ffd7-6b0c-484b-acda-110f0ffb5a5d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.296482 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" podUID="4697ffd7-6b0c-484b-acda-110f0ffb5a5d" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.337042 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.337171 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9ml2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-ql9fl_openstack(4abed129-55b7-4d35-99e4-d8817c7dc39e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.340269 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" podUID="4abed129-55b7-4d35-99e4-d8817c7dc39e" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.371668 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.372080 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kkl9p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-8rp4w_openstack(dcbd3e2c-a1f8-461a-b30a-6e819a696cdb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:22:50 crc kubenswrapper[4735]: E1122 08:22:50.373598 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" podUID="dcbd3e2c-a1f8-461a-b30a-6e819a696cdb" Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.056600 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c"] Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.068252 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z5csg"] Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.188379 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef","Type":"ContainerStarted","Data":"eab3e7622255f5be2dd8b16d3e815b740d0f1bd96d034966e563aabff982038d"} Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.188444 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.189725 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" event={"ID":"36da6203-0695-45d5-b361-01267bd61a1d","Type":"ContainerStarted","Data":"cd4cbdc704a75782419d31b44dd0678527700d60eff6042b061ceb1a1cc75e6e"} Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.192551 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0c715895-af70-4e8f-96c4-c7064e45c2c9","Type":"ContainerStarted","Data":"68d7644156f728d99feb0d88c12b4723fdc97dd9a15216f2bbed37903f2863cb"} Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.199250 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1","Type":"ContainerStarted","Data":"303f445e725205be0f6215d1ed6cf91e4576ffc3f3625125cfbbae186c5d0bd3"} Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.208398 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg" event={"ID":"d99e6f50-a748-4ae8-b1b0-9df901b37201","Type":"ContainerStarted","Data":"0963d826d90fa147dfdd968262ab943e299793bb16b2a574db487a0a9d5497e0"} Nov 22 08:22:51 crc kubenswrapper[4735]: E1122 08:22:51.213027 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" podUID="dcbd3e2c-a1f8-461a-b30a-6e819a696cdb" Nov 22 08:22:51 crc kubenswrapper[4735]: E1122 08:22:51.213083 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" podUID="4abed129-55b7-4d35-99e4-d8817c7dc39e" Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.222136 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=9.867200223 podStartE2EDuration="21.222114019s" podCreationTimestamp="2025-11-22 08:22:30 +0000 UTC" firstStartedPulling="2025-11-22 08:22:38.949235909 +0000 UTC m=+1180.553574514" lastFinishedPulling="2025-11-22 08:22:50.304149705 +0000 UTC m=+1191.908488310" observedRunningTime="2025-11-22 08:22:51.207130523 +0000 UTC m=+1192.811469138" watchObservedRunningTime="2025-11-22 08:22:51.222114019 +0000 UTC m=+1192.826452624" Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.300379 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.300415 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.318611 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5df59bb7db-gjtgp"] Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.894019 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 08:22:51 crc kubenswrapper[4735]: I1122 08:22:51.998768 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.007668 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.185808 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-dns-svc\") pod \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.185885 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz29v\" (UniqueName: \"kubernetes.io/projected/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-kube-api-access-gz29v\") pod \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.185940 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29dql\" (UniqueName: \"kubernetes.io/projected/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-kube-api-access-29dql\") pod \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\" (UID: \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\") " Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.186075 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-config\") pod \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\" (UID: \"8f9d9209-ed3c-49fa-83cb-bdaa27a65487\") " Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.186147 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-config\") pod \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\" (UID: \"4697ffd7-6b0c-484b-acda-110f0ffb5a5d\") " Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.186374 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4697ffd7-6b0c-484b-acda-110f0ffb5a5d" (UID: "4697ffd7-6b0c-484b-acda-110f0ffb5a5d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.186978 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-config" (OuterVolumeSpecName: "config") pod "4697ffd7-6b0c-484b-acda-110f0ffb5a5d" (UID: "4697ffd7-6b0c-484b-acda-110f0ffb5a5d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.187167 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.187433 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-config" (OuterVolumeSpecName: "config") pod "8f9d9209-ed3c-49fa-83cb-bdaa27a65487" (UID: "8f9d9209-ed3c-49fa-83cb-bdaa27a65487"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.191826 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-kube-api-access-29dql" (OuterVolumeSpecName: "kube-api-access-29dql") pod "8f9d9209-ed3c-49fa-83cb-bdaa27a65487" (UID: "8f9d9209-ed3c-49fa-83cb-bdaa27a65487"). InnerVolumeSpecName "kube-api-access-29dql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.193191 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-kube-api-access-gz29v" (OuterVolumeSpecName: "kube-api-access-gz29v") pod "4697ffd7-6b0c-484b-acda-110f0ffb5a5d" (UID: "4697ffd7-6b0c-484b-acda-110f0ffb5a5d"). InnerVolumeSpecName "kube-api-access-gz29v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.225469 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4c7911f8-389a-43fe-a011-dcc65ccbd7f2","Type":"ContainerStarted","Data":"21fdf04e69138814f021a78e2b82930f926643c1eacc99f95d0811e5978b4e63"} Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.227152 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.227166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9lbg4" event={"ID":"8f9d9209-ed3c-49fa-83cb-bdaa27a65487","Type":"ContainerDied","Data":"9b4a60522141c197454ce890a0ae2ffdf7519b5013448dde734f04af44229f64"} Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.229849 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" event={"ID":"4697ffd7-6b0c-484b-acda-110f0ffb5a5d","Type":"ContainerDied","Data":"4dd6ead2feab6adc705bdb213ae56cda7a4884c15e644036e739ff233204179b"} Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.230027 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-55wh7" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.234303 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10","Type":"ContainerStarted","Data":"639b9585adb4796e0e70cf58fd13241ab15ccbd2b150f54e3680a54b17b7030f"} Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.242044 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5df59bb7db-gjtgp" event={"ID":"0e0587d6-2a11-4bb9-9be8-ac139a2ff912","Type":"ContainerStarted","Data":"aa63bcef332df77030cf3a03ebaed46f1020edc75bde5a2dc87976fdc2d131ab"} Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.242084 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5df59bb7db-gjtgp" event={"ID":"0e0587d6-2a11-4bb9-9be8-ac139a2ff912","Type":"ContainerStarted","Data":"6046fb078df527d3ed6b4114959d806531658dc84382c559cbc35cdc4112ca87"} Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.245199 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"16a0218b-2a6b-47f5-832f-2d8da9f950f3","Type":"ContainerStarted","Data":"95a812981f253ce98ee2c3bd8e0df2da7cee96f083de4c16330ad2ae6bb49eac"} Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.247554 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerStarted","Data":"85f8d0995934a219cb503ddae2705f2d58f00c911193c039df90ff7cc690d342"} Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.293926 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.293956 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz29v\" (UniqueName: \"kubernetes.io/projected/4697ffd7-6b0c-484b-acda-110f0ffb5a5d-kube-api-access-gz29v\") on node \"crc\" DevicePath \"\"" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.293966 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29dql\" (UniqueName: \"kubernetes.io/projected/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-kube-api-access-29dql\") on node \"crc\" DevicePath \"\"" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.293978 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9d9209-ed3c-49fa-83cb-bdaa27a65487-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.317260 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5df59bb7db-gjtgp" podStartSLOduration=19.317236706 podStartE2EDuration="19.317236706s" podCreationTimestamp="2025-11-22 08:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:22:52.28641848 +0000 UTC m=+1193.890757095" watchObservedRunningTime="2025-11-22 08:22:52.317236706 +0000 UTC m=+1193.921575311" Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.349600 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9lbg4"] Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.357393 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9lbg4"] Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.376769 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55wh7"] Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.386047 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55wh7"] Nov 22 08:22:52 crc kubenswrapper[4735]: I1122 08:22:52.716767 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 08:22:53 crc kubenswrapper[4735]: I1122 08:22:53.335762 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4697ffd7-6b0c-484b-acda-110f0ffb5a5d" path="/var/lib/kubelet/pods/4697ffd7-6b0c-484b-acda-110f0ffb5a5d/volumes" Nov 22 08:22:53 crc kubenswrapper[4735]: I1122 08:22:53.336151 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f9d9209-ed3c-49fa-83cb-bdaa27a65487" path="/var/lib/kubelet/pods/8f9d9209-ed3c-49fa-83cb-bdaa27a65487/volumes" Nov 22 08:22:53 crc kubenswrapper[4735]: I1122 08:22:53.381341 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7mhwj"] Nov 22 08:22:54 crc kubenswrapper[4735]: I1122 08:22:54.010996 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:54 crc kubenswrapper[4735]: I1122 08:22:54.011609 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:54 crc kubenswrapper[4735]: I1122 08:22:54.015173 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:54 crc kubenswrapper[4735]: I1122 08:22:54.359597 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"135fc91d-a8bd-489a-a7ae-3df9db9af6cd","Type":"ContainerStarted","Data":"eb8bbef37579955f4dba6b9836ee1396ebc22b96c2ecef390b6aae8eafd3d4df"} Nov 22 08:22:54 crc kubenswrapper[4735]: I1122 08:22:54.361487 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7mhwj" event={"ID":"fee31945-9fca-4693-9218-97bb4513cee3","Type":"ContainerStarted","Data":"386f4bb155017da273992cf212ba581488d39b201db7539251f98c4eb838ee9e"} Nov 22 08:22:54 crc kubenswrapper[4735]: I1122 08:22:54.367489 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5df59bb7db-gjtgp" Nov 22 08:22:54 crc kubenswrapper[4735]: I1122 08:22:54.423785 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-6d7db65f59-c579x"] Nov 22 08:22:55 crc kubenswrapper[4735]: I1122 08:22:55.380956 4735 generic.go:334] "Generic (PLEG): container finished" podID="0c715895-af70-4e8f-96c4-c7064e45c2c9" containerID="68d7644156f728d99feb0d88c12b4723fdc97dd9a15216f2bbed37903f2863cb" exitCode=0 Nov 22 08:22:55 crc kubenswrapper[4735]: I1122 08:22:55.381027 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0c715895-af70-4e8f-96c4-c7064e45c2c9","Type":"ContainerDied","Data":"68d7644156f728d99feb0d88c12b4723fdc97dd9a15216f2bbed37903f2863cb"} Nov 22 08:22:55 crc kubenswrapper[4735]: I1122 08:22:55.383831 4735 generic.go:334] "Generic (PLEG): container finished" podID="5eaebfb3-8113-41d6-9a3d-77c8194ad4d1" containerID="303f445e725205be0f6215d1ed6cf91e4576ffc3f3625125cfbbae186c5d0bd3" exitCode=0 Nov 22 08:22:55 crc kubenswrapper[4735]: I1122 08:22:55.384299 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1","Type":"ContainerDied","Data":"303f445e725205be0f6215d1ed6cf91e4576ffc3f3625125cfbbae186c5d0bd3"} Nov 22 08:22:55 crc kubenswrapper[4735]: I1122 08:22:55.965445 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.415137 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"16a0218b-2a6b-47f5-832f-2d8da9f950f3","Type":"ContainerStarted","Data":"0ed542edda862f50ea5bbad6f788191455dc4bf429f623214f7a440f2b1d7234"} Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.417064 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0c715895-af70-4e8f-96c4-c7064e45c2c9","Type":"ContainerStarted","Data":"4aac1a13518555a7d9fd816635f29ea84f5b3def44c298a8de58c56dd13e6acb"} Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.418359 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4c7911f8-389a-43fe-a011-dcc65ccbd7f2","Type":"ContainerStarted","Data":"786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9"} Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.418559 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.420003 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5eaebfb3-8113-41d6-9a3d-77c8194ad4d1","Type":"ContainerStarted","Data":"53cfdf61356c8fc2b7ed7a7497583120aeaacc199bedd34964972195f4e2273d"} Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.421159 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg" event={"ID":"d99e6f50-a748-4ae8-b1b0-9df901b37201","Type":"ContainerStarted","Data":"a9e4a4c7088df1d63a8bf4f94f0b97806f4d862ad864ab28bedb0dc79760a504"} Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.421209 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-z5csg" Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.422307 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7mhwj" event={"ID":"fee31945-9fca-4693-9218-97bb4513cee3","Type":"ContainerStarted","Data":"c8c02e35fb67effe91f64b58dcf50f06f9b997539df2872e900e7ee5ba12cb2a"} Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.423400 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" event={"ID":"36da6203-0695-45d5-b361-01267bd61a1d","Type":"ContainerStarted","Data":"5a115940b022675757d3c4da6b6ec1440c800313aee05b224487d1d46bf7d872"} Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.424511 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"135fc91d-a8bd-489a-a7ae-3df9db9af6cd","Type":"ContainerStarted","Data":"e136ec7d5cf478c5a343492a2fb9a21ae446dcb5d638a31de5c6d87e52f2fe2b"} Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.439997 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.04956533 podStartE2EDuration="29.439973041s" podCreationTimestamp="2025-11-22 08:22:29 +0000 UTC" firstStartedPulling="2025-11-22 08:22:38.951243735 +0000 UTC m=+1180.555582340" lastFinishedPulling="2025-11-22 08:22:50.341651446 +0000 UTC m=+1191.945990051" observedRunningTime="2025-11-22 08:22:58.435312391 +0000 UTC m=+1200.039651006" watchObservedRunningTime="2025-11-22 08:22:58.439973041 +0000 UTC m=+1200.044311656" Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.458187 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-2dl4c" podStartSLOduration=19.973966614 podStartE2EDuration="25.458166826s" podCreationTimestamp="2025-11-22 08:22:33 +0000 UTC" firstStartedPulling="2025-11-22 08:22:51.072122402 +0000 UTC m=+1192.676461017" lastFinishedPulling="2025-11-22 08:22:56.556322624 +0000 UTC m=+1198.160661229" observedRunningTime="2025-11-22 08:22:58.456238653 +0000 UTC m=+1200.060577258" watchObservedRunningTime="2025-11-22 08:22:58.458166826 +0000 UTC m=+1200.062505431" Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.518945 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=11.419750839 podStartE2EDuration="31.518921234s" podCreationTimestamp="2025-11-22 08:22:27 +0000 UTC" firstStartedPulling="2025-11-22 08:22:30.213887857 +0000 UTC m=+1171.818226462" lastFinishedPulling="2025-11-22 08:22:50.313058242 +0000 UTC m=+1191.917396857" observedRunningTime="2025-11-22 08:22:58.507849497 +0000 UTC m=+1200.112188112" watchObservedRunningTime="2025-11-22 08:22:58.518921234 +0000 UTC m=+1200.123259839" Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.527705 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-z5csg" podStartSLOduration=17.607233714 podStartE2EDuration="23.527688528s" podCreationTimestamp="2025-11-22 08:22:35 +0000 UTC" firstStartedPulling="2025-11-22 08:22:51.076667078 +0000 UTC m=+1192.681005683" lastFinishedPulling="2025-11-22 08:22:56.997121862 +0000 UTC m=+1198.601460497" observedRunningTime="2025-11-22 08:22:58.524935411 +0000 UTC m=+1200.129274036" watchObservedRunningTime="2025-11-22 08:22:58.527688528 +0000 UTC m=+1200.132027133" Nov 22 08:22:58 crc kubenswrapper[4735]: I1122 08:22:58.555411 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=20.354797771 podStartE2EDuration="26.555394748s" podCreationTimestamp="2025-11-22 08:22:32 +0000 UTC" firstStartedPulling="2025-11-22 08:22:51.301728612 +0000 UTC m=+1192.906067217" lastFinishedPulling="2025-11-22 08:22:57.502325579 +0000 UTC m=+1199.106664194" observedRunningTime="2025-11-22 08:22:58.540654068 +0000 UTC m=+1200.144992683" watchObservedRunningTime="2025-11-22 08:22:58.555394748 +0000 UTC m=+1200.159733353" Nov 22 08:22:59 crc kubenswrapper[4735]: I1122 08:22:59.400307 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 22 08:22:59 crc kubenswrapper[4735]: I1122 08:22:59.400616 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 22 08:22:59 crc kubenswrapper[4735]: I1122 08:22:59.435874 4735 generic.go:334] "Generic (PLEG): container finished" podID="fee31945-9fca-4693-9218-97bb4513cee3" containerID="c8c02e35fb67effe91f64b58dcf50f06f9b997539df2872e900e7ee5ba12cb2a" exitCode=0 Nov 22 08:22:59 crc kubenswrapper[4735]: I1122 08:22:59.437510 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7mhwj" event={"ID":"fee31945-9fca-4693-9218-97bb4513cee3","Type":"ContainerDied","Data":"c8c02e35fb67effe91f64b58dcf50f06f9b997539df2872e900e7ee5ba12cb2a"} Nov 22 08:23:00 crc kubenswrapper[4735]: I1122 08:23:00.674554 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 22 08:23:00 crc kubenswrapper[4735]: I1122 08:23:00.674803 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 22 08:23:01 crc kubenswrapper[4735]: I1122 08:23:01.452161 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerStarted","Data":"ce33ea0327b60ba451a8c4b8f72bad7f53c1270c3e4e6bb4012f9b3c0c94f30c"} Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.463031 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"135fc91d-a8bd-489a-a7ae-3df9db9af6cd","Type":"ContainerStarted","Data":"a9e12fa344c51375bb91451cb7713f0c17de3b7c1a6666b9e6b5b739e7326e62"} Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.472161 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"16a0218b-2a6b-47f5-832f-2d8da9f950f3","Type":"ContainerStarted","Data":"7bc42948a31a7b511244a246f4041212af08347269cad6c52b79c728d01d1d29"} Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.474977 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7mhwj" event={"ID":"fee31945-9fca-4693-9218-97bb4513cee3","Type":"ContainerStarted","Data":"3f7636e406822ddcd470051a96472bc524edc732322c25abb97d1bb3767205ed"} Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.486931 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=18.939621403 podStartE2EDuration="27.486915682s" podCreationTimestamp="2025-11-22 08:22:35 +0000 UTC" firstStartedPulling="2025-11-22 08:22:53.499783401 +0000 UTC m=+1195.104121996" lastFinishedPulling="2025-11-22 08:23:02.04707766 +0000 UTC m=+1203.651416275" observedRunningTime="2025-11-22 08:23:02.485821671 +0000 UTC m=+1204.090160286" watchObservedRunningTime="2025-11-22 08:23:02.486915682 +0000 UTC m=+1204.091254287" Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.518596 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.40519254 podStartE2EDuration="24.518576921s" podCreationTimestamp="2025-11-22 08:22:38 +0000 UTC" firstStartedPulling="2025-11-22 08:22:51.90995167 +0000 UTC m=+1193.514290285" lastFinishedPulling="2025-11-22 08:23:02.023336061 +0000 UTC m=+1203.627674666" observedRunningTime="2025-11-22 08:23:02.505726293 +0000 UTC m=+1204.110064898" watchObservedRunningTime="2025-11-22 08:23:02.518576921 +0000 UTC m=+1204.122915526" Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.849812 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ql9fl"] Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.858096 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-h4cpd"] Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.859884 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:02 crc kubenswrapper[4735]: I1122 08:23:02.878304 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-h4cpd"] Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.062388 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5lns\" (UniqueName: \"kubernetes.io/projected/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-kube-api-access-f5lns\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.062772 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.062845 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-config\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.165959 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5lns\" (UniqueName: \"kubernetes.io/projected/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-kube-api-access-f5lns\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.166035 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.166090 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-config\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.166992 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-config\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.176158 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.255092 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5lns\" (UniqueName: \"kubernetes.io/projected/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-kube-api-access-f5lns\") pod \"dnsmasq-dns-7cb5889db5-h4cpd\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.355425 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.487982 4735 generic.go:334] "Generic (PLEG): container finished" podID="4abed129-55b7-4d35-99e4-d8817c7dc39e" containerID="958c5390f7728dbaca066336abbd1e4ae974e7352011985257dd26a85a110b4f" exitCode=0 Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.488060 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" event={"ID":"4abed129-55b7-4d35-99e4-d8817c7dc39e","Type":"ContainerDied","Data":"958c5390f7728dbaca066336abbd1e4ae974e7352011985257dd26a85a110b4f"} Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.495344 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.501595 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7mhwj" event={"ID":"fee31945-9fca-4693-9218-97bb4513cee3","Type":"ContainerStarted","Data":"2e0d7257a5deaf27e10e13384f7fa36015bb2f9b4bd3334b821de8e68e3d3339"} Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.502432 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.502544 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.575117 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7mhwj" podStartSLOduration=25.087105145 podStartE2EDuration="28.575096215s" podCreationTimestamp="2025-11-22 08:22:35 +0000 UTC" firstStartedPulling="2025-11-22 08:22:53.507719902 +0000 UTC m=+1195.112058507" lastFinishedPulling="2025-11-22 08:22:56.995710932 +0000 UTC m=+1198.600049577" observedRunningTime="2025-11-22 08:23:03.570888898 +0000 UTC m=+1205.175227503" watchObservedRunningTime="2025-11-22 08:23:03.575096215 +0000 UTC m=+1205.179434820" Nov 22 08:23:03 crc kubenswrapper[4735]: E1122 08:23:03.816879 4735 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 22 08:23:03 crc kubenswrapper[4735]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 22 08:23:03 crc kubenswrapper[4735]: > podSandboxID="3471102029aa054e8554ef597d7a9718139712ae51aa6c22ca4174aad9c9db08" Nov 22 08:23:03 crc kubenswrapper[4735]: E1122 08:23:03.818241 4735 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 08:23:03 crc kubenswrapper[4735]: init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kkl9p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-8rp4w_openstack(dcbd3e2c-a1f8-461a-b30a-6e819a696cdb): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 22 08:23:03 crc kubenswrapper[4735]: > logger="UnhandledError" Nov 22 08:23:03 crc kubenswrapper[4735]: E1122 08:23:03.823811 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" podUID="dcbd3e2c-a1f8-461a-b30a-6e819a696cdb" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.844440 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.867662 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.880605 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.884412 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.884418 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-ckb9k" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.884577 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.886448 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.903951 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 22 08:23:03 crc kubenswrapper[4735]: I1122 08:23:03.929977 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.002013 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/49e1ddee-0d04-4d19-9c87-e438e5e70e70-cache\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.002116 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/49e1ddee-0d04-4d19-9c87-e438e5e70e70-lock\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.002145 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.002189 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.002205 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzn2q\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-kube-api-access-gzn2q\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.010124 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.060778 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.097752 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.109471 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.109523 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.109550 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzn2q\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-kube-api-access-gzn2q\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.109910 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/49e1ddee-0d04-4d19-9c87-e438e5e70e70-cache\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.110003 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/49e1ddee-0d04-4d19-9c87-e438e5e70e70-lock\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.110778 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.111775 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/49e1ddee-0d04-4d19-9c87-e438e5e70e70-lock\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.112356 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/49e1ddee-0d04-4d19-9c87-e438e5e70e70-cache\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: E1122 08:23:04.112448 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 22 08:23:04 crc kubenswrapper[4735]: E1122 08:23:04.112492 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 22 08:23:04 crc kubenswrapper[4735]: E1122 08:23:04.112528 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift podName:49e1ddee-0d04-4d19-9c87-e438e5e70e70 nodeName:}" failed. No retries permitted until 2025-11-22 08:23:04.612511977 +0000 UTC m=+1206.216850582 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift") pod "swift-storage-0" (UID: "49e1ddee-0d04-4d19-9c87-e438e5e70e70") : configmap "swift-ring-files" not found Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.137549 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzn2q\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-kube-api-access-gzn2q\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.172588 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.178219 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.211410 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-config\") pod \"4abed129-55b7-4d35-99e4-d8817c7dc39e\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.211492 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ml2v\" (UniqueName: \"kubernetes.io/projected/4abed129-55b7-4d35-99e4-d8817c7dc39e-kube-api-access-9ml2v\") pod \"4abed129-55b7-4d35-99e4-d8817c7dc39e\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.211557 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-dns-svc\") pod \"4abed129-55b7-4d35-99e4-d8817c7dc39e\" (UID: \"4abed129-55b7-4d35-99e4-d8817c7dc39e\") " Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.220085 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4abed129-55b7-4d35-99e4-d8817c7dc39e-kube-api-access-9ml2v" (OuterVolumeSpecName: "kube-api-access-9ml2v") pod "4abed129-55b7-4d35-99e4-d8817c7dc39e" (UID: "4abed129-55b7-4d35-99e4-d8817c7dc39e"). InnerVolumeSpecName "kube-api-access-9ml2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.238894 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4abed129-55b7-4d35-99e4-d8817c7dc39e" (UID: "4abed129-55b7-4d35-99e4-d8817c7dc39e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.245991 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-config" (OuterVolumeSpecName: "config") pod "4abed129-55b7-4d35-99e4-d8817c7dc39e" (UID: "4abed129-55b7-4d35-99e4-d8817c7dc39e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.261423 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-h4cpd"] Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.314590 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.314625 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ml2v\" (UniqueName: \"kubernetes.io/projected/4abed129-55b7-4d35-99e4-d8817c7dc39e-kube-api-access-9ml2v\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.314634 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4abed129-55b7-4d35-99e4-d8817c7dc39e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.516007 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" event={"ID":"b1d0d45a-218b-40ff-81fc-77c0f955ec0f","Type":"ContainerStarted","Data":"bdc955733f07a0d54ecd660eda51944d68cb8b6e34b4186615f89545ec0c46d8"} Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.517860 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5ee793a0-c437-47a3-b05d-369a1730d6b1","Type":"ContainerStarted","Data":"b6b68c78ecac956d2a32b20d00cbfd5998f52fb623cb15919609639792d34e37"} Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.519440 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" event={"ID":"4abed129-55b7-4d35-99e4-d8817c7dc39e","Type":"ContainerDied","Data":"0900d3d2c8f811738df1aa2faecaad36597ec988dcf983141f8553d496b156e7"} Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.519516 4735 scope.go:117] "RemoveContainer" containerID="958c5390f7728dbaca066336abbd1e4ae974e7352011985257dd26a85a110b4f" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.519822 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.519925 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ql9fl" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.520525 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.573219 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.631998 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:04 crc kubenswrapper[4735]: E1122 08:23:04.632699 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 22 08:23:04 crc kubenswrapper[4735]: E1122 08:23:04.632727 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 22 08:23:04 crc kubenswrapper[4735]: E1122 08:23:04.632790 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift podName:49e1ddee-0d04-4d19-9c87-e438e5e70e70 nodeName:}" failed. No retries permitted until 2025-11-22 08:23:05.632762152 +0000 UTC m=+1207.237100757 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift") pod "swift-storage-0" (UID: "49e1ddee-0d04-4d19-9c87-e438e5e70e70") : configmap "swift-ring-files" not found Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.635202 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.665083 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ql9fl"] Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.681739 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ql9fl"] Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.908952 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8rp4w"] Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.939485 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vps6w"] Nov 22 08:23:04 crc kubenswrapper[4735]: E1122 08:23:04.941426 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4abed129-55b7-4d35-99e4-d8817c7dc39e" containerName="init" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.941619 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4abed129-55b7-4d35-99e4-d8817c7dc39e" containerName="init" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.941969 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4abed129-55b7-4d35-99e4-d8817c7dc39e" containerName="init" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.954707 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.957534 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.975009 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-4tx8r"] Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.976360 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.982410 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 22 08:23:04 crc kubenswrapper[4735]: I1122 08:23:04.995338 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vps6w"] Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.014942 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4tx8r"] Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.044102 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/451e7b5a-64f4-4f0a-8a44-32a86375b765-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.044508 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.044652 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/451e7b5a-64f4-4f0a-8a44-32a86375b765-combined-ca-bundle\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.044797 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsfgz\" (UniqueName: \"kubernetes.io/projected/451e7b5a-64f4-4f0a-8a44-32a86375b765-kube-api-access-jsfgz\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.044926 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/451e7b5a-64f4-4f0a-8a44-32a86375b765-ovs-rundir\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.045016 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.045091 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/451e7b5a-64f4-4f0a-8a44-32a86375b765-ovn-rundir\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.045192 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fk77\" (UniqueName: \"kubernetes.io/projected/05727804-ca8c-48b6-adaf-e86a36263c3c-kube-api-access-9fk77\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.045286 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-config\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.045413 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/451e7b5a-64f4-4f0a-8a44-32a86375b765-config\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.071649 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-h4cpd"] Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.104534 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-nrccs"] Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.106362 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.109614 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.127662 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nrccs"] Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.147779 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z75m5\" (UniqueName: \"kubernetes.io/projected/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-kube-api-access-z75m5\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.147912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-config\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.148911 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-config\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.149096 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/451e7b5a-64f4-4f0a-8a44-32a86375b765-config\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.149774 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/451e7b5a-64f4-4f0a-8a44-32a86375b765-config\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.149914 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150265 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-dns-svc\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150410 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-config\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150499 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/451e7b5a-64f4-4f0a-8a44-32a86375b765-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150619 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150732 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/451e7b5a-64f4-4f0a-8a44-32a86375b765-combined-ca-bundle\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150777 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsfgz\" (UniqueName: \"kubernetes.io/projected/451e7b5a-64f4-4f0a-8a44-32a86375b765-kube-api-access-jsfgz\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150801 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/451e7b5a-64f4-4f0a-8a44-32a86375b765-ovs-rundir\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150817 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150843 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/451e7b5a-64f4-4f0a-8a44-32a86375b765-ovn-rundir\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150859 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.150910 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fk77\" (UniqueName: \"kubernetes.io/projected/05727804-ca8c-48b6-adaf-e86a36263c3c-kube-api-access-9fk77\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.151248 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.152233 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/451e7b5a-64f4-4f0a-8a44-32a86375b765-ovn-rundir\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.152443 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/451e7b5a-64f4-4f0a-8a44-32a86375b765-ovs-rundir\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.152905 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.180407 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/451e7b5a-64f4-4f0a-8a44-32a86375b765-combined-ca-bundle\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.180890 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/451e7b5a-64f4-4f0a-8a44-32a86375b765-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.183030 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.192559 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.199661 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fk77\" (UniqueName: \"kubernetes.io/projected/05727804-ca8c-48b6-adaf-e86a36263c3c-kube-api-access-9fk77\") pod \"dnsmasq-dns-6c89d5d749-vps6w\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.205077 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.205385 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.205652 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-xcdxv" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.205892 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.205972 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsfgz\" (UniqueName: \"kubernetes.io/projected/451e7b5a-64f4-4f0a-8a44-32a86375b765-kube-api-access-jsfgz\") pod \"ovn-controller-metrics-4tx8r\" (UID: \"451e7b5a-64f4-4f0a-8a44-32a86375b765\") " pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.221602 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.253700 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264026 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-dns-svc\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264100 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264140 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-config\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264181 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264205 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8l8f\" (UniqueName: \"kubernetes.io/projected/3ebf2f32-a55d-4844-85b3-bb09482d04e1-kube-api-access-g8l8f\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264233 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ebf2f32-a55d-4844-85b3-bb09482d04e1-scripts\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264262 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ebf2f32-a55d-4844-85b3-bb09482d04e1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264306 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264354 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264401 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z75m5\" (UniqueName: \"kubernetes.io/projected/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-kube-api-access-z75m5\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.264442 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ebf2f32-a55d-4844-85b3-bb09482d04e1-config\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.254810 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.273397 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.273985 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-dns-svc\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.274321 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-config\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.283928 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4abed129-55b7-4d35-99e4-d8817c7dc39e" path="/var/lib/kubelet/pods/4abed129-55b7-4d35-99e4-d8817c7dc39e/volumes" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.302352 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z75m5\" (UniqueName: \"kubernetes.io/projected/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-kube-api-access-z75m5\") pod \"dnsmasq-dns-698758b865-nrccs\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.309586 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.320965 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4tx8r" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.377082 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8l8f\" (UniqueName: \"kubernetes.io/projected/3ebf2f32-a55d-4844-85b3-bb09482d04e1-kube-api-access-g8l8f\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.377153 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ebf2f32-a55d-4844-85b3-bb09482d04e1-scripts\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.377199 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ebf2f32-a55d-4844-85b3-bb09482d04e1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.377293 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.377447 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ebf2f32-a55d-4844-85b3-bb09482d04e1-config\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.377617 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.377704 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.378871 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ebf2f32-a55d-4844-85b3-bb09482d04e1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.380585 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ebf2f32-a55d-4844-85b3-bb09482d04e1-scripts\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.384653 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.387210 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ebf2f32-a55d-4844-85b3-bb09482d04e1-config\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.387951 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.390572 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ebf2f32-a55d-4844-85b3-bb09482d04e1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.412260 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8l8f\" (UniqueName: \"kubernetes.io/projected/3ebf2f32-a55d-4844-85b3-bb09482d04e1-kube-api-access-g8l8f\") pod \"ovn-northd-0\" (UID: \"3ebf2f32-a55d-4844-85b3-bb09482d04e1\") " pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.465334 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.475146 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.541899 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" event={"ID":"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb","Type":"ContainerDied","Data":"3471102029aa054e8554ef597d7a9718139712ae51aa6c22ca4174aad9c9db08"} Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.541985 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8rp4w" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.545576 4735 generic.go:334] "Generic (PLEG): container finished" podID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" containerID="441ae411e853adca0269a63cc5728189a4b2512284cff06f71f23994508bd48d" exitCode=0 Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.546469 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" event={"ID":"b1d0d45a-218b-40ff-81fc-77c0f955ec0f","Type":"ContainerDied","Data":"441ae411e853adca0269a63cc5728189a4b2512284cff06f71f23994508bd48d"} Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.570283 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.582730 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkl9p\" (UniqueName: \"kubernetes.io/projected/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-kube-api-access-kkl9p\") pod \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.583011 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config\") pod \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.583072 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-dns-svc\") pod \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.593006 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-kube-api-access-kkl9p" (OuterVolumeSpecName: "kube-api-access-kkl9p") pod "dcbd3e2c-a1f8-461a-b30a-6e819a696cdb" (UID: "dcbd3e2c-a1f8-461a-b30a-6e819a696cdb"). InnerVolumeSpecName "kube-api-access-kkl9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:05 crc kubenswrapper[4735]: E1122 08:23:05.617511 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config podName:dcbd3e2c-a1f8-461a-b30a-6e819a696cdb nodeName:}" failed. No retries permitted until 2025-11-22 08:23:06.11744103 +0000 UTC m=+1207.721779635 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config") pod "dcbd3e2c-a1f8-461a-b30a-6e819a696cdb" (UID: "dcbd3e2c-a1f8-461a-b30a-6e819a696cdb") : error deleting /var/lib/kubelet/pods/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb/volume-subpaths: remove /var/lib/kubelet/pods/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb/volume-subpaths: no such file or directory Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.617924 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dcbd3e2c-a1f8-461a-b30a-6e819a696cdb" (UID: "dcbd3e2c-a1f8-461a-b30a-6e819a696cdb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.690911 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.690998 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkl9p\" (UniqueName: \"kubernetes.io/projected/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-kube-api-access-kkl9p\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:05 crc kubenswrapper[4735]: I1122 08:23:05.691009 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:05 crc kubenswrapper[4735]: E1122 08:23:05.693350 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 22 08:23:05 crc kubenswrapper[4735]: E1122 08:23:05.693378 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 22 08:23:05 crc kubenswrapper[4735]: E1122 08:23:05.693410 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift podName:49e1ddee-0d04-4d19-9c87-e438e5e70e70 nodeName:}" failed. No retries permitted until 2025-11-22 08:23:07.69339787 +0000 UTC m=+1209.297736475 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift") pod "swift-storage-0" (UID: "49e1ddee-0d04-4d19-9c87-e438e5e70e70") : configmap "swift-ring-files" not found Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.174284 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4tx8r"] Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.203121 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config\") pod \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\" (UID: \"dcbd3e2c-a1f8-461a-b30a-6e819a696cdb\") " Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.203906 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config" (OuterVolumeSpecName: "config") pod "dcbd3e2c-a1f8-461a-b30a-6e819a696cdb" (UID: "dcbd3e2c-a1f8-461a-b30a-6e819a696cdb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.241800 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vps6w"] Nov 22 08:23:06 crc kubenswrapper[4735]: W1122 08:23:06.271012 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05727804_ca8c_48b6_adaf_e86a36263c3c.slice/crio-4786e9e74aee4f7c722b6d9355d4f7a45be387a39b6e9fc2cd284dbfedb2a372 WatchSource:0}: Error finding container 4786e9e74aee4f7c722b6d9355d4f7a45be387a39b6e9fc2cd284dbfedb2a372: Status 404 returned error can't find the container with id 4786e9e74aee4f7c722b6d9355d4f7a45be387a39b6e9fc2cd284dbfedb2a372 Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.305764 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.341433 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nrccs"] Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.349727 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 22 08:23:06 crc kubenswrapper[4735]: W1122 08:23:06.360624 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ee474f0_0a9f_4c80_bacb_6140ac9c571c.slice/crio-9d184d168ea858104727f4f2818845eb756e8ac24f6563405e7c9699e8622904 WatchSource:0}: Error finding container 9d184d168ea858104727f4f2818845eb756e8ac24f6563405e7c9699e8622904: Status 404 returned error can't find the container with id 9d184d168ea858104727f4f2818845eb756e8ac24f6563405e7c9699e8622904 Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.542540 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8rp4w"] Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.553642 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8rp4w"] Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.564725 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" event={"ID":"05727804-ca8c-48b6-adaf-e86a36263c3c","Type":"ContainerStarted","Data":"4786e9e74aee4f7c722b6d9355d4f7a45be387a39b6e9fc2cd284dbfedb2a372"} Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.568536 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4tx8r" event={"ID":"451e7b5a-64f4-4f0a-8a44-32a86375b765","Type":"ContainerStarted","Data":"31f69c51f62615c9d636b96f0024d706b04adb03c20ca32c44c2038cf71cd79b"} Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.571703 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" event={"ID":"b1d0d45a-218b-40ff-81fc-77c0f955ec0f","Type":"ContainerStarted","Data":"cf2255b08732c9fab20f0760e3bd6b8e8625a66616e358a3dcd69a26addb2d33"} Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.571976 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" podUID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" containerName="dnsmasq-dns" containerID="cri-o://cf2255b08732c9fab20f0760e3bd6b8e8625a66616e358a3dcd69a26addb2d33" gracePeriod=10 Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.572074 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.579078 4735 generic.go:334] "Generic (PLEG): container finished" podID="bea30a30-b625-4511-87da-f26e5495a5f9" containerID="ce33ea0327b60ba451a8c4b8f72bad7f53c1270c3e4e6bb4012f9b3c0c94f30c" exitCode=0 Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.579190 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerDied","Data":"ce33ea0327b60ba451a8c4b8f72bad7f53c1270c3e4e6bb4012f9b3c0c94f30c"} Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.588518 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ebf2f32-a55d-4844-85b3-bb09482d04e1","Type":"ContainerStarted","Data":"fb95b8a90eb969c5572f8833b528ede96f9038e8b74e0796bccc1cffd2d33217"} Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.596513 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nrccs" event={"ID":"3ee474f0-0a9f-4c80-bacb-6140ac9c571c","Type":"ContainerStarted","Data":"9d184d168ea858104727f4f2818845eb756e8ac24f6563405e7c9699e8622904"} Nov 22 08:23:06 crc kubenswrapper[4735]: I1122 08:23:06.598955 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" podStartSLOduration=4.598938729 podStartE2EDuration="4.598938729s" podCreationTimestamp="2025-11-22 08:23:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:06.595629197 +0000 UTC m=+1208.199967792" watchObservedRunningTime="2025-11-22 08:23:06.598938729 +0000 UTC m=+1208.203277334" Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.276151 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcbd3e2c-a1f8-461a-b30a-6e819a696cdb" path="/var/lib/kubelet/pods/dcbd3e2c-a1f8-461a-b30a-6e819a696cdb/volumes" Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.608030 4735 generic.go:334] "Generic (PLEG): container finished" podID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerID="ec0ad732d773a8ecbde959d4860287f82f5db5c8095a3fa21e3a39061b5a6023" exitCode=0 Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.608113 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nrccs" event={"ID":"3ee474f0-0a9f-4c80-bacb-6140ac9c571c","Type":"ContainerDied","Data":"ec0ad732d773a8ecbde959d4860287f82f5db5c8095a3fa21e3a39061b5a6023"} Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.609915 4735 generic.go:334] "Generic (PLEG): container finished" podID="05727804-ca8c-48b6-adaf-e86a36263c3c" containerID="4ceba5e68a13c3d1a5af6b8a3587abf671726c5c98bf992784ed5c5fa99ea41f" exitCode=0 Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.609961 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" event={"ID":"05727804-ca8c-48b6-adaf-e86a36263c3c","Type":"ContainerDied","Data":"4ceba5e68a13c3d1a5af6b8a3587abf671726c5c98bf992784ed5c5fa99ea41f"} Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.612400 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4tx8r" event={"ID":"451e7b5a-64f4-4f0a-8a44-32a86375b765","Type":"ContainerStarted","Data":"933930b6ac2c42c32f28f1ee2c4421eb34cb718a59c6b5c65cc637e722f6fc61"} Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.614724 4735 generic.go:334] "Generic (PLEG): container finished" podID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" containerID="cf2255b08732c9fab20f0760e3bd6b8e8625a66616e358a3dcd69a26addb2d33" exitCode=0 Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.614772 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" event={"ID":"b1d0d45a-218b-40ff-81fc-77c0f955ec0f","Type":"ContainerDied","Data":"cf2255b08732c9fab20f0760e3bd6b8e8625a66616e358a3dcd69a26addb2d33"} Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.667140 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-4tx8r" podStartSLOduration=3.667118778 podStartE2EDuration="3.667118778s" podCreationTimestamp="2025-11-22 08:23:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:07.666346426 +0000 UTC m=+1209.270685031" watchObservedRunningTime="2025-11-22 08:23:07.667118778 +0000 UTC m=+1209.271457383" Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.745519 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:07 crc kubenswrapper[4735]: E1122 08:23:07.745723 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 22 08:23:07 crc kubenswrapper[4735]: E1122 08:23:07.745957 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 22 08:23:07 crc kubenswrapper[4735]: E1122 08:23:07.746061 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift podName:49e1ddee-0d04-4d19-9c87-e438e5e70e70 nodeName:}" failed. No retries permitted until 2025-11-22 08:23:11.746045621 +0000 UTC m=+1213.350384226 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift") pod "swift-storage-0" (UID: "49e1ddee-0d04-4d19-9c87-e438e5e70e70") : configmap "swift-ring-files" not found Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.897338 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-hvj22"] Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.898984 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.900756 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.900913 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.908033 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 22 08:23:07 crc kubenswrapper[4735]: I1122 08:23:07.912280 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hvj22"] Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.056325 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxfbr\" (UniqueName: \"kubernetes.io/projected/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-kube-api-access-xxfbr\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.056373 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-combined-ca-bundle\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.056391 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-swiftconf\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.056516 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-ring-data-devices\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.056563 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-scripts\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.056658 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-dispersionconf\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.056751 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-etc-swift\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.159334 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxfbr\" (UniqueName: \"kubernetes.io/projected/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-kube-api-access-xxfbr\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.159404 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-combined-ca-bundle\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.159436 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-swiftconf\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.159551 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-ring-data-devices\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.159614 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-scripts\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.159663 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-dispersionconf\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.159900 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-etc-swift\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.160342 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-etc-swift\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.161567 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-scripts\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.162407 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-ring-data-devices\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.168551 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-combined-ca-bundle\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.176516 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-swiftconf\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.176453 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-dispersionconf\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.177854 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxfbr\" (UniqueName: \"kubernetes.io/projected/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-kube-api-access-xxfbr\") pod \"swift-ring-rebalance-hvj22\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.261509 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:08 crc kubenswrapper[4735]: I1122 08:23:08.745647 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hvj22"] Nov 22 08:23:09 crc kubenswrapper[4735]: I1122 08:23:09.637758 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hvj22" event={"ID":"abf32873-0945-47b6-8f4f-ed1b18dc6eb1","Type":"ContainerStarted","Data":"3a89ceb6f55c9e6e788f9307665f29ca82a875ad03c7abbe4b70b81cd7ac2681"} Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.216092 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.323833 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.652574 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.670236 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" event={"ID":"05727804-ca8c-48b6-adaf-e86a36263c3c","Type":"ContainerStarted","Data":"328ee9ef102328cae7327a50f8ee34bd45ad68bf2f30298f9f3aafa182c6cb00"} Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.670376 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.673558 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" event={"ID":"b1d0d45a-218b-40ff-81fc-77c0f955ec0f","Type":"ContainerDied","Data":"bdc955733f07a0d54ecd660eda51944d68cb8b6e34b4186615f89545ec0c46d8"} Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.673611 4735 scope.go:117] "RemoveContainer" containerID="cf2255b08732c9fab20f0760e3bd6b8e8625a66616e358a3dcd69a26addb2d33" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.673837 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-h4cpd" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.681255 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nrccs" event={"ID":"3ee474f0-0a9f-4c80-bacb-6140ac9c571c","Type":"ContainerStarted","Data":"ced057972c34b33ab51b58b80d8b4d0fb8a52b95f5cf6bf5837b966ff7a17164"} Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.684877 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.709572 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" podStartSLOduration=7.709555593 podStartE2EDuration="7.709555593s" podCreationTimestamp="2025-11-22 08:23:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:11.703548376 +0000 UTC m=+1213.307886991" watchObservedRunningTime="2025-11-22 08:23:11.709555593 +0000 UTC m=+1213.313894198" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.743112 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-nrccs" podStartSLOduration=6.743091565 podStartE2EDuration="6.743091565s" podCreationTimestamp="2025-11-22 08:23:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:11.739358901 +0000 UTC m=+1213.343697506" watchObservedRunningTime="2025-11-22 08:23:11.743091565 +0000 UTC m=+1213.347430170" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.770066 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-dns-svc\") pod \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.770147 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-config\") pod \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.770394 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5lns\" (UniqueName: \"kubernetes.io/projected/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-kube-api-access-f5lns\") pod \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\" (UID: \"b1d0d45a-218b-40ff-81fc-77c0f955ec0f\") " Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.770767 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:11 crc kubenswrapper[4735]: E1122 08:23:11.771017 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 22 08:23:11 crc kubenswrapper[4735]: E1122 08:23:11.771037 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 22 08:23:11 crc kubenswrapper[4735]: E1122 08:23:11.771085 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift podName:49e1ddee-0d04-4d19-9c87-e438e5e70e70 nodeName:}" failed. No retries permitted until 2025-11-22 08:23:19.771069432 +0000 UTC m=+1221.375408037 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift") pod "swift-storage-0" (UID: "49e1ddee-0d04-4d19-9c87-e438e5e70e70") : configmap "swift-ring-files" not found Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.803729 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-kube-api-access-f5lns" (OuterVolumeSpecName: "kube-api-access-f5lns") pod "b1d0d45a-218b-40ff-81fc-77c0f955ec0f" (UID: "b1d0d45a-218b-40ff-81fc-77c0f955ec0f"). InnerVolumeSpecName "kube-api-access-f5lns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.874839 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5lns\" (UniqueName: \"kubernetes.io/projected/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-kube-api-access-f5lns\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.902257 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-config" (OuterVolumeSpecName: "config") pod "b1d0d45a-218b-40ff-81fc-77c0f955ec0f" (UID: "b1d0d45a-218b-40ff-81fc-77c0f955ec0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.928451 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b1d0d45a-218b-40ff-81fc-77c0f955ec0f" (UID: "b1d0d45a-218b-40ff-81fc-77c0f955ec0f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.988757 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:11 crc kubenswrapper[4735]: I1122 08:23:11.988789 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d0d45a-218b-40ff-81fc-77c0f955ec0f-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.069155 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-h4cpd"] Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.076510 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-h4cpd"] Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.351075 4735 scope.go:117] "RemoveContainer" containerID="441ae411e853adca0269a63cc5728189a4b2512284cff06f71f23994508bd48d" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.877234 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-2f1c-account-create-f4v9s"] Nov 22 08:23:12 crc kubenswrapper[4735]: E1122 08:23:12.878840 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" containerName="init" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.878950 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" containerName="init" Nov 22 08:23:12 crc kubenswrapper[4735]: E1122 08:23:12.878999 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" containerName="dnsmasq-dns" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.879008 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" containerName="dnsmasq-dns" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.879270 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" containerName="dnsmasq-dns" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.880491 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.885563 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.904200 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-2f1c-account-create-f4v9s"] Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.908743 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.920150 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-nfpfq"] Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.923890 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:12 crc kubenswrapper[4735]: I1122 08:23:12.945331 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-nfpfq"] Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.014056 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t22bh\" (UniqueName: \"kubernetes.io/projected/6e8014b5-baf9-4531-a275-2df3ba0c2af1-kube-api-access-t22bh\") pod \"mysqld-exporter-openstack-db-create-nfpfq\" (UID: \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\") " pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.014215 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7957d0-44ed-4753-a100-253797cb1be9-operator-scripts\") pod \"mysqld-exporter-2f1c-account-create-f4v9s\" (UID: \"4e7957d0-44ed-4753-a100-253797cb1be9\") " pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.014259 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf84p\" (UniqueName: \"kubernetes.io/projected/4e7957d0-44ed-4753-a100-253797cb1be9-kube-api-access-gf84p\") pod \"mysqld-exporter-2f1c-account-create-f4v9s\" (UID: \"4e7957d0-44ed-4753-a100-253797cb1be9\") " pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.014303 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e8014b5-baf9-4531-a275-2df3ba0c2af1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-nfpfq\" (UID: \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\") " pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.116202 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t22bh\" (UniqueName: \"kubernetes.io/projected/6e8014b5-baf9-4531-a275-2df3ba0c2af1-kube-api-access-t22bh\") pod \"mysqld-exporter-openstack-db-create-nfpfq\" (UID: \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\") " pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.116357 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7957d0-44ed-4753-a100-253797cb1be9-operator-scripts\") pod \"mysqld-exporter-2f1c-account-create-f4v9s\" (UID: \"4e7957d0-44ed-4753-a100-253797cb1be9\") " pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.116403 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf84p\" (UniqueName: \"kubernetes.io/projected/4e7957d0-44ed-4753-a100-253797cb1be9-kube-api-access-gf84p\") pod \"mysqld-exporter-2f1c-account-create-f4v9s\" (UID: \"4e7957d0-44ed-4753-a100-253797cb1be9\") " pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.116450 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e8014b5-baf9-4531-a275-2df3ba0c2af1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-nfpfq\" (UID: \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\") " pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.122281 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e8014b5-baf9-4531-a275-2df3ba0c2af1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-nfpfq\" (UID: \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\") " pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.122899 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7957d0-44ed-4753-a100-253797cb1be9-operator-scripts\") pod \"mysqld-exporter-2f1c-account-create-f4v9s\" (UID: \"4e7957d0-44ed-4753-a100-253797cb1be9\") " pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.173634 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf84p\" (UniqueName: \"kubernetes.io/projected/4e7957d0-44ed-4753-a100-253797cb1be9-kube-api-access-gf84p\") pod \"mysqld-exporter-2f1c-account-create-f4v9s\" (UID: \"4e7957d0-44ed-4753-a100-253797cb1be9\") " pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.226327 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.263538 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t22bh\" (UniqueName: \"kubernetes.io/projected/6e8014b5-baf9-4531-a275-2df3ba0c2af1-kube-api-access-t22bh\") pod \"mysqld-exporter-openstack-db-create-nfpfq\" (UID: \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\") " pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.274860 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1d0d45a-218b-40ff-81fc-77c0f955ec0f" path="/var/lib/kubelet/pods/b1d0d45a-218b-40ff-81fc-77c0f955ec0f/volumes" Nov 22 08:23:13 crc kubenswrapper[4735]: I1122 08:23:13.559888 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.125020 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ztsfq"] Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.127002 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.170376 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ztsfq"] Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.227531 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c517-account-create-gg4qg"] Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.232037 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.234112 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.254326 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c517-account-create-gg4qg"] Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.286601 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98303611-37f9-4fee-89fb-c4b58d51ff2f-operator-scripts\") pod \"glance-db-create-ztsfq\" (UID: \"98303611-37f9-4fee-89fb-c4b58d51ff2f\") " pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.287035 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg2wb\" (UniqueName: \"kubernetes.io/projected/98303611-37f9-4fee-89fb-c4b58d51ff2f-kube-api-access-rg2wb\") pod \"glance-db-create-ztsfq\" (UID: \"98303611-37f9-4fee-89fb-c4b58d51ff2f\") " pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.389241 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77979ac0-cbb3-4785-9bd5-14cfb29749f3-operator-scripts\") pod \"glance-c517-account-create-gg4qg\" (UID: \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\") " pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.389434 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg2wb\" (UniqueName: \"kubernetes.io/projected/98303611-37f9-4fee-89fb-c4b58d51ff2f-kube-api-access-rg2wb\") pod \"glance-db-create-ztsfq\" (UID: \"98303611-37f9-4fee-89fb-c4b58d51ff2f\") " pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.389637 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98303611-37f9-4fee-89fb-c4b58d51ff2f-operator-scripts\") pod \"glance-db-create-ztsfq\" (UID: \"98303611-37f9-4fee-89fb-c4b58d51ff2f\") " pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.390656 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98303611-37f9-4fee-89fb-c4b58d51ff2f-operator-scripts\") pod \"glance-db-create-ztsfq\" (UID: \"98303611-37f9-4fee-89fb-c4b58d51ff2f\") " pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.391582 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmjmd\" (UniqueName: \"kubernetes.io/projected/77979ac0-cbb3-4785-9bd5-14cfb29749f3-kube-api-access-pmjmd\") pod \"glance-c517-account-create-gg4qg\" (UID: \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\") " pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.413064 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg2wb\" (UniqueName: \"kubernetes.io/projected/98303611-37f9-4fee-89fb-c4b58d51ff2f-kube-api-access-rg2wb\") pod \"glance-db-create-ztsfq\" (UID: \"98303611-37f9-4fee-89fb-c4b58d51ff2f\") " pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.492549 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.493567 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmjmd\" (UniqueName: \"kubernetes.io/projected/77979ac0-cbb3-4785-9bd5-14cfb29749f3-kube-api-access-pmjmd\") pod \"glance-c517-account-create-gg4qg\" (UID: \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\") " pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.493762 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77979ac0-cbb3-4785-9bd5-14cfb29749f3-operator-scripts\") pod \"glance-c517-account-create-gg4qg\" (UID: \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\") " pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.494548 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77979ac0-cbb3-4785-9bd5-14cfb29749f3-operator-scripts\") pod \"glance-c517-account-create-gg4qg\" (UID: \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\") " pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.512302 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmjmd\" (UniqueName: \"kubernetes.io/projected/77979ac0-cbb3-4785-9bd5-14cfb29749f3-kube-api-access-pmjmd\") pod \"glance-c517-account-create-gg4qg\" (UID: \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\") " pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:16 crc kubenswrapper[4735]: I1122 08:23:16.548959 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:19 crc kubenswrapper[4735]: I1122 08:23:19.478160 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-6d7db65f59-c579x" podUID="847def28-51ed-4e0d-8629-79c707aebb60" containerName="console" containerID="cri-o://7c333a1cc524ddd1fe128428db18e84333cdb9a5bf8934630e4854f6454a3c0c" gracePeriod=15 Nov 22 08:23:19 crc kubenswrapper[4735]: I1122 08:23:19.780301 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:19 crc kubenswrapper[4735]: E1122 08:23:19.780795 4735 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 22 08:23:19 crc kubenswrapper[4735]: E1122 08:23:19.782441 4735 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 22 08:23:19 crc kubenswrapper[4735]: E1122 08:23:19.782643 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift podName:49e1ddee-0d04-4d19-9c87-e438e5e70e70 nodeName:}" failed. No retries permitted until 2025-11-22 08:23:35.782621504 +0000 UTC m=+1237.386960109 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift") pod "swift-storage-0" (UID: "49e1ddee-0d04-4d19-9c87-e438e5e70e70") : configmap "swift-ring-files" not found Nov 22 08:23:19 crc kubenswrapper[4735]: I1122 08:23:19.797565 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6d7db65f59-c579x_847def28-51ed-4e0d-8629-79c707aebb60/console/0.log" Nov 22 08:23:19 crc kubenswrapper[4735]: I1122 08:23:19.797616 4735 generic.go:334] "Generic (PLEG): container finished" podID="847def28-51ed-4e0d-8629-79c707aebb60" containerID="7c333a1cc524ddd1fe128428db18e84333cdb9a5bf8934630e4854f6454a3c0c" exitCode=2 Nov 22 08:23:19 crc kubenswrapper[4735]: I1122 08:23:19.797672 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d7db65f59-c579x" event={"ID":"847def28-51ed-4e0d-8629-79c707aebb60","Type":"ContainerDied","Data":"7c333a1cc524ddd1fe128428db18e84333cdb9a5bf8934630e4854f6454a3c0c"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.226684 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6d7db65f59-c579x_847def28-51ed-4e0d-8629-79c707aebb60/console/0.log" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.227005 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.311899 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.394498 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-trusted-ca-bundle\") pod \"847def28-51ed-4e0d-8629-79c707aebb60\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.394672 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-oauth-config\") pod \"847def28-51ed-4e0d-8629-79c707aebb60\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.394717 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-service-ca\") pod \"847def28-51ed-4e0d-8629-79c707aebb60\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.394753 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-serving-cert\") pod \"847def28-51ed-4e0d-8629-79c707aebb60\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.394823 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-oauth-serving-cert\") pod \"847def28-51ed-4e0d-8629-79c707aebb60\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.394867 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-console-config\") pod \"847def28-51ed-4e0d-8629-79c707aebb60\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.394921 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqtcp\" (UniqueName: \"kubernetes.io/projected/847def28-51ed-4e0d-8629-79c707aebb60-kube-api-access-fqtcp\") pod \"847def28-51ed-4e0d-8629-79c707aebb60\" (UID: \"847def28-51ed-4e0d-8629-79c707aebb60\") " Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.397032 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "847def28-51ed-4e0d-8629-79c707aebb60" (UID: "847def28-51ed-4e0d-8629-79c707aebb60"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.398717 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-service-ca" (OuterVolumeSpecName: "service-ca") pod "847def28-51ed-4e0d-8629-79c707aebb60" (UID: "847def28-51ed-4e0d-8629-79c707aebb60"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.405771 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "847def28-51ed-4e0d-8629-79c707aebb60" (UID: "847def28-51ed-4e0d-8629-79c707aebb60"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.405928 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "847def28-51ed-4e0d-8629-79c707aebb60" (UID: "847def28-51ed-4e0d-8629-79c707aebb60"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.406047 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-console-config" (OuterVolumeSpecName: "console-config") pod "847def28-51ed-4e0d-8629-79c707aebb60" (UID: "847def28-51ed-4e0d-8629-79c707aebb60"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.407219 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "847def28-51ed-4e0d-8629-79c707aebb60" (UID: "847def28-51ed-4e0d-8629-79c707aebb60"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.415744 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/847def28-51ed-4e0d-8629-79c707aebb60-kube-api-access-fqtcp" (OuterVolumeSpecName: "kube-api-access-fqtcp") pod "847def28-51ed-4e0d-8629-79c707aebb60" (UID: "847def28-51ed-4e0d-8629-79c707aebb60"). InnerVolumeSpecName "kube-api-access-fqtcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.453762 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2n7mj"] Nov 22 08:23:20 crc kubenswrapper[4735]: E1122 08:23:20.454163 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847def28-51ed-4e0d-8629-79c707aebb60" containerName="console" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.454179 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="847def28-51ed-4e0d-8629-79c707aebb60" containerName="console" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.454364 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="847def28-51ed-4e0d-8629-79c707aebb60" containerName="console" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.455023 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.468469 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.469926 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2n7mj"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.498336 4735 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.498371 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqtcp\" (UniqueName: \"kubernetes.io/projected/847def28-51ed-4e0d-8629-79c707aebb60-kube-api-access-fqtcp\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.498381 4735 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.498390 4735 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.498399 4735 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.498408 4735 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/847def28-51ed-4e0d-8629-79c707aebb60-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.498416 4735 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/847def28-51ed-4e0d-8629-79c707aebb60-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.569832 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ztsfq"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.570334 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vps6w"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.592629 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f25a-account-create-rg9dx"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.594001 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.596587 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.602298 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z287l\" (UniqueName: \"kubernetes.io/projected/cf0ea764-b639-4959-8a27-e2f674be3329-kube-api-access-z287l\") pod \"keystone-db-create-2n7mj\" (UID: \"cf0ea764-b639-4959-8a27-e2f674be3329\") " pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.602350 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0ea764-b639-4959-8a27-e2f674be3329-operator-scripts\") pod \"keystone-db-create-2n7mj\" (UID: \"cf0ea764-b639-4959-8a27-e2f674be3329\") " pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.607135 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c517-account-create-gg4qg"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.633917 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-nfpfq"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.650592 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-2f1c-account-create-f4v9s"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.665185 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f25a-account-create-rg9dx"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.704015 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tlpv\" (UniqueName: \"kubernetes.io/projected/659efe09-eb81-4668-a68f-fc72b7bc43d1-kube-api-access-6tlpv\") pod \"keystone-f25a-account-create-rg9dx\" (UID: \"659efe09-eb81-4668-a68f-fc72b7bc43d1\") " pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.704185 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z287l\" (UniqueName: \"kubernetes.io/projected/cf0ea764-b639-4959-8a27-e2f674be3329-kube-api-access-z287l\") pod \"keystone-db-create-2n7mj\" (UID: \"cf0ea764-b639-4959-8a27-e2f674be3329\") " pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.704214 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0ea764-b639-4959-8a27-e2f674be3329-operator-scripts\") pod \"keystone-db-create-2n7mj\" (UID: \"cf0ea764-b639-4959-8a27-e2f674be3329\") " pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.704240 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/659efe09-eb81-4668-a68f-fc72b7bc43d1-operator-scripts\") pod \"keystone-f25a-account-create-rg9dx\" (UID: \"659efe09-eb81-4668-a68f-fc72b7bc43d1\") " pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.705201 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0ea764-b639-4959-8a27-e2f674be3329-operator-scripts\") pod \"keystone-db-create-2n7mj\" (UID: \"cf0ea764-b639-4959-8a27-e2f674be3329\") " pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.734995 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z287l\" (UniqueName: \"kubernetes.io/projected/cf0ea764-b639-4959-8a27-e2f674be3329-kube-api-access-z287l\") pod \"keystone-db-create-2n7mj\" (UID: \"cf0ea764-b639-4959-8a27-e2f674be3329\") " pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.776592 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.806850 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/659efe09-eb81-4668-a68f-fc72b7bc43d1-operator-scripts\") pod \"keystone-f25a-account-create-rg9dx\" (UID: \"659efe09-eb81-4668-a68f-fc72b7bc43d1\") " pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.806963 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tlpv\" (UniqueName: \"kubernetes.io/projected/659efe09-eb81-4668-a68f-fc72b7bc43d1-kube-api-access-6tlpv\") pod \"keystone-f25a-account-create-rg9dx\" (UID: \"659efe09-eb81-4668-a68f-fc72b7bc43d1\") " pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.808043 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/659efe09-eb81-4668-a68f-fc72b7bc43d1-operator-scripts\") pod \"keystone-f25a-account-create-rg9dx\" (UID: \"659efe09-eb81-4668-a68f-fc72b7bc43d1\") " pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.812056 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-dmgvb"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.816333 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.842242 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-dmgvb"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.842282 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c517-account-create-gg4qg" event={"ID":"77979ac0-cbb3-4785-9bd5-14cfb29749f3","Type":"ContainerStarted","Data":"a84dc53b42e22c5c5de7c87e82d1e80035b89850d69a872ef257b860f1bbab6a"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.845642 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tlpv\" (UniqueName: \"kubernetes.io/projected/659efe09-eb81-4668-a68f-fc72b7bc43d1-kube-api-access-6tlpv\") pod \"keystone-f25a-account-create-rg9dx\" (UID: \"659efe09-eb81-4668-a68f-fc72b7bc43d1\") " pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.854864 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ebf2f32-a55d-4844-85b3-bb09482d04e1","Type":"ContainerStarted","Data":"df038256a9750ccf9d617d16ffcaac772a2612bfa208b10378d31fec5846eb7d"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.854922 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ebf2f32-a55d-4844-85b3-bb09482d04e1","Type":"ContainerStarted","Data":"0fc7c60c511b28ca47b0a45eb919a9d598238c6fb3ec27717ff5dc42da925a32"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.855214 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.861771 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ztsfq" event={"ID":"98303611-37f9-4fee-89fb-c4b58d51ff2f","Type":"ContainerStarted","Data":"c5a4480f978fee54aead122316c043b3a0a69ad365c49443c0e055fa67dacfcf"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.888083 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hvj22" event={"ID":"abf32873-0945-47b6-8f4f-ed1b18dc6eb1","Type":"ContainerStarted","Data":"0ff683dd6063093793ecd3f43038e6e3e4b89d6469aaaefc88bf285ffadc0a6a"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.900634 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-20d8-account-create-crjxr"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.902096 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.905779 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.906531 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" event={"ID":"4e7957d0-44ed-4753-a100-253797cb1be9","Type":"ContainerStarted","Data":"052a7dac6e92065c41f8b425939543dabad99a972b44791188b63338b37da2aa"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.911825 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqt8w\" (UniqueName: \"kubernetes.io/projected/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-kube-api-access-qqt8w\") pod \"placement-db-create-dmgvb\" (UID: \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\") " pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.911946 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-operator-scripts\") pod \"placement-db-create-dmgvb\" (UID: \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\") " pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.919378 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6d7db65f59-c579x_847def28-51ed-4e0d-8629-79c707aebb60/console/0.log" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.919472 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d7db65f59-c579x" event={"ID":"847def28-51ed-4e0d-8629-79c707aebb60","Type":"ContainerDied","Data":"bb7cd2942fc23837733d386015a0bbadf60a31edacfa3c0c9ea151810bcc79ed"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.919514 4735 scope.go:117] "RemoveContainer" containerID="7c333a1cc524ddd1fe128428db18e84333cdb9a5bf8934630e4854f6454a3c0c" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.919640 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d7db65f59-c579x" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.923015 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" event={"ID":"6e8014b5-baf9-4531-a275-2df3ba0c2af1","Type":"ContainerStarted","Data":"99a603e2a85a21d64e6a8e6da47115d7139ea6a869994e6b0f162f0fe8d4d80d"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.929378 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-20d8-account-create-crjxr"] Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.935414 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=6.716583297 podStartE2EDuration="15.935394553s" podCreationTimestamp="2025-11-22 08:23:05 +0000 UTC" firstStartedPulling="2025-11-22 08:23:06.373608119 +0000 UTC m=+1207.977946724" lastFinishedPulling="2025-11-22 08:23:15.592419375 +0000 UTC m=+1217.196757980" observedRunningTime="2025-11-22 08:23:20.900143734 +0000 UTC m=+1222.504482359" watchObservedRunningTime="2025-11-22 08:23:20.935394553 +0000 UTC m=+1222.539733158" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.947152 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" podUID="05727804-ca8c-48b6-adaf-e86a36263c3c" containerName="dnsmasq-dns" containerID="cri-o://328ee9ef102328cae7327a50f8ee34bd45ad68bf2f30298f9f3aafa182c6cb00" gracePeriod=10 Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.947620 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerStarted","Data":"b3c9fc6ea8757266331eabc4a4429d89d80fb829b60d8538119de99da812b8ae"} Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.958572 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-hvj22" podStartSLOduration=3.021728537 podStartE2EDuration="13.958549887s" podCreationTimestamp="2025-11-22 08:23:07 +0000 UTC" firstStartedPulling="2025-11-22 08:23:08.759008684 +0000 UTC m=+1210.363347289" lastFinishedPulling="2025-11-22 08:23:19.695830034 +0000 UTC m=+1221.300168639" observedRunningTime="2025-11-22 08:23:20.929027157 +0000 UTC m=+1222.533365762" watchObservedRunningTime="2025-11-22 08:23:20.958549887 +0000 UTC m=+1222.562888492" Nov 22 08:23:20 crc kubenswrapper[4735]: I1122 08:23:20.985250 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.016255 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-operator-scripts\") pod \"placement-db-create-dmgvb\" (UID: \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\") " pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.016317 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzckl\" (UniqueName: \"kubernetes.io/projected/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-kube-api-access-zzckl\") pod \"placement-20d8-account-create-crjxr\" (UID: \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\") " pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.016649 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-operator-scripts\") pod \"placement-20d8-account-create-crjxr\" (UID: \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\") " pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.016873 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqt8w\" (UniqueName: \"kubernetes.io/projected/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-kube-api-access-qqt8w\") pod \"placement-db-create-dmgvb\" (UID: \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\") " pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.017238 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-operator-scripts\") pod \"placement-db-create-dmgvb\" (UID: \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\") " pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.036404 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqt8w\" (UniqueName: \"kubernetes.io/projected/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-kube-api-access-qqt8w\") pod \"placement-db-create-dmgvb\" (UID: \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\") " pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.109521 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.119309 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzckl\" (UniqueName: \"kubernetes.io/projected/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-kube-api-access-zzckl\") pod \"placement-20d8-account-create-crjxr\" (UID: \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\") " pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.119554 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-operator-scripts\") pod \"placement-20d8-account-create-crjxr\" (UID: \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\") " pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.120537 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-operator-scripts\") pod \"placement-20d8-account-create-crjxr\" (UID: \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\") " pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.149314 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-6d7db65f59-c579x"] Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.168103 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-6d7db65f59-c579x"] Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.173735 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzckl\" (UniqueName: \"kubernetes.io/projected/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-kube-api-access-zzckl\") pod \"placement-20d8-account-create-crjxr\" (UID: \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\") " pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.315632 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="847def28-51ed-4e0d-8629-79c707aebb60" path="/var/lib/kubelet/pods/847def28-51ed-4e0d-8629-79c707aebb60/volumes" Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.406188 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2n7mj"] Nov 22 08:23:21 crc kubenswrapper[4735]: I1122 08:23:21.423480 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:21 crc kubenswrapper[4735]: W1122 08:23:21.448907 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf0ea764_b639_4959_8a27_e2f674be3329.slice/crio-9ced673df6504baf62400eb1bf5fbe637e3f7623781f4dfeca97a4459e0f98d4 WatchSource:0}: Error finding container 9ced673df6504baf62400eb1bf5fbe637e3f7623781f4dfeca97a4459e0f98d4: Status 404 returned error can't find the container with id 9ced673df6504baf62400eb1bf5fbe637e3f7623781f4dfeca97a4459e0f98d4 Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.756946 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f25a-account-create-rg9dx"] Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.905373 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-dmgvb"] Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.955235 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2n7mj" event={"ID":"cf0ea764-b639-4959-8a27-e2f674be3329","Type":"ContainerStarted","Data":"9ced673df6504baf62400eb1bf5fbe637e3f7623781f4dfeca97a4459e0f98d4"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.957099 4735 generic.go:334] "Generic (PLEG): container finished" podID="6e8014b5-baf9-4531-a275-2df3ba0c2af1" containerID="1edd19d6822f1ac35e33e83fc3835af064231a438259df80b686b83e4d16e74c" exitCode=0 Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.957534 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" event={"ID":"6e8014b5-baf9-4531-a275-2df3ba0c2af1","Type":"ContainerDied","Data":"1edd19d6822f1ac35e33e83fc3835af064231a438259df80b686b83e4d16e74c"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.958842 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f25a-account-create-rg9dx" event={"ID":"659efe09-eb81-4668-a68f-fc72b7bc43d1","Type":"ContainerStarted","Data":"614b89b3383419337c9604506293dc47428b789bd63bea663eb10431193e5966"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.961081 4735 generic.go:334] "Generic (PLEG): container finished" podID="77979ac0-cbb3-4785-9bd5-14cfb29749f3" containerID="27d9ba4ee2fcd61435f2e708644df7ef6daba680a65e417058ef81995c759113" exitCode=0 Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.961160 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c517-account-create-gg4qg" event={"ID":"77979ac0-cbb3-4785-9bd5-14cfb29749f3","Type":"ContainerDied","Data":"27d9ba4ee2fcd61435f2e708644df7ef6daba680a65e417058ef81995c759113"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.963005 4735 generic.go:334] "Generic (PLEG): container finished" podID="98303611-37f9-4fee-89fb-c4b58d51ff2f" containerID="07bc0ff0794476d96188f8305cd08785cb95998803e5ce2165a534bb8edd86b7" exitCode=0 Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.963076 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ztsfq" event={"ID":"98303611-37f9-4fee-89fb-c4b58d51ff2f","Type":"ContainerDied","Data":"07bc0ff0794476d96188f8305cd08785cb95998803e5ce2165a534bb8edd86b7"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.966943 4735 generic.go:334] "Generic (PLEG): container finished" podID="05727804-ca8c-48b6-adaf-e86a36263c3c" containerID="328ee9ef102328cae7327a50f8ee34bd45ad68bf2f30298f9f3aafa182c6cb00" exitCode=0 Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.967063 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" event={"ID":"05727804-ca8c-48b6-adaf-e86a36263c3c","Type":"ContainerDied","Data":"328ee9ef102328cae7327a50f8ee34bd45ad68bf2f30298f9f3aafa182c6cb00"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.967110 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" event={"ID":"05727804-ca8c-48b6-adaf-e86a36263c3c","Type":"ContainerDied","Data":"4786e9e74aee4f7c722b6d9355d4f7a45be387a39b6e9fc2cd284dbfedb2a372"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.967128 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4786e9e74aee4f7c722b6d9355d4f7a45be387a39b6e9fc2cd284dbfedb2a372" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.973200 4735 generic.go:334] "Generic (PLEG): container finished" podID="4e7957d0-44ed-4753-a100-253797cb1be9" containerID="8708cea40636f856f6bf29e50ac1ba1056fb28cd103ea3cf009c4a29eaebb3d9" exitCode=0 Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:21.973405 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" event={"ID":"4e7957d0-44ed-4753-a100-253797cb1be9","Type":"ContainerDied","Data":"8708cea40636f856f6bf29e50ac1ba1056fb28cd103ea3cf009c4a29eaebb3d9"} Nov 22 08:23:22 crc kubenswrapper[4735]: W1122 08:23:21.980714 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f4ab0e4_1b2f_4c1a_8414_d85c38509a64.slice/crio-8414ecf2510e6deecb481777cb0b87f65cdd219281e18d3480095c0221bc8d82 WatchSource:0}: Error finding container 8414ecf2510e6deecb481777cb0b87f65cdd219281e18d3480095c0221bc8d82: Status 404 returned error can't find the container with id 8414ecf2510e6deecb481777cb0b87f65cdd219281e18d3480095c0221bc8d82 Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.038668 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.153448 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-20d8-account-create-crjxr"] Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.158585 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-ovsdbserver-sb\") pod \"05727804-ca8c-48b6-adaf-e86a36263c3c\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.158731 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-config\") pod \"05727804-ca8c-48b6-adaf-e86a36263c3c\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.158816 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-dns-svc\") pod \"05727804-ca8c-48b6-adaf-e86a36263c3c\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.158904 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fk77\" (UniqueName: \"kubernetes.io/projected/05727804-ca8c-48b6-adaf-e86a36263c3c-kube-api-access-9fk77\") pod \"05727804-ca8c-48b6-adaf-e86a36263c3c\" (UID: \"05727804-ca8c-48b6-adaf-e86a36263c3c\") " Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.165848 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05727804-ca8c-48b6-adaf-e86a36263c3c-kube-api-access-9fk77" (OuterVolumeSpecName: "kube-api-access-9fk77") pod "05727804-ca8c-48b6-adaf-e86a36263c3c" (UID: "05727804-ca8c-48b6-adaf-e86a36263c3c"). InnerVolumeSpecName "kube-api-access-9fk77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.262048 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fk77\" (UniqueName: \"kubernetes.io/projected/05727804-ca8c-48b6-adaf-e86a36263c3c-kube-api-access-9fk77\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.428375 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "05727804-ca8c-48b6-adaf-e86a36263c3c" (UID: "05727804-ca8c-48b6-adaf-e86a36263c3c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.467079 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.552193 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-config" (OuterVolumeSpecName: "config") pod "05727804-ca8c-48b6-adaf-e86a36263c3c" (UID: "05727804-ca8c-48b6-adaf-e86a36263c3c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.554849 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "05727804-ca8c-48b6-adaf-e86a36263c3c" (UID: "05727804-ca8c-48b6-adaf-e86a36263c3c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.568499 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.568521 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05727804-ca8c-48b6-adaf-e86a36263c3c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.774639 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:23:22 crc kubenswrapper[4735]: E1122 08:23:22.784531 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf0ea764_b639_4959_8a27_e2f674be3329.slice/crio-51567db7aa746db0f673c4bd05a1613a8503972d79f08a7a09e0acfc13a81172.scope\": RecentStats: unable to find data in memory cache]" Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.987809 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-20d8-account-create-crjxr" event={"ID":"c5e4fc11-5ae0-4336-bc26-f07ca5230b48","Type":"ContainerStarted","Data":"a5bcee5f440f8e179c2ad85fc45661bf74347ae9960f818494e1ba614685722d"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.987860 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-20d8-account-create-crjxr" event={"ID":"c5e4fc11-5ae0-4336-bc26-f07ca5230b48","Type":"ContainerStarted","Data":"8c4a8dc3c2d5ed0c190d8202f28f065db34ab962c34ba47bb3832daaeb248098"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.990383 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dmgvb" event={"ID":"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64","Type":"ContainerStarted","Data":"406a5f6bf3dcb697458cd3bd8db577b52baa9653d96a6c8c308d44f0a457e919"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.990413 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dmgvb" event={"ID":"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64","Type":"ContainerStarted","Data":"8414ecf2510e6deecb481777cb0b87f65cdd219281e18d3480095c0221bc8d82"} Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.992760 4735 generic.go:334] "Generic (PLEG): container finished" podID="cf0ea764-b639-4959-8a27-e2f674be3329" containerID="51567db7aa746db0f673c4bd05a1613a8503972d79f08a7a09e0acfc13a81172" exitCode=0 Nov 22 08:23:22 crc kubenswrapper[4735]: I1122 08:23:22.992861 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2n7mj" event={"ID":"cf0ea764-b639-4959-8a27-e2f674be3329","Type":"ContainerDied","Data":"51567db7aa746db0f673c4bd05a1613a8503972d79f08a7a09e0acfc13a81172"} Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.006000 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerStarted","Data":"33d4f40ef552d8b1c5dbed8c3b3d4e62283ddfc16a37f2822441144dcf24ecc5"} Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.010850 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-20d8-account-create-crjxr" podStartSLOduration=3.010831318 podStartE2EDuration="3.010831318s" podCreationTimestamp="2025-11-22 08:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:23.007945327 +0000 UTC m=+1224.612283942" watchObservedRunningTime="2025-11-22 08:23:23.010831318 +0000 UTC m=+1224.615169923" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.011330 4735 generic.go:334] "Generic (PLEG): container finished" podID="659efe09-eb81-4668-a68f-fc72b7bc43d1" containerID="e787fd4b999aade029525ca31150f0ca69d81e74d922b10588eea746b41a8001" exitCode=0 Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.012994 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f25a-account-create-rg9dx" event={"ID":"659efe09-eb81-4668-a68f-fc72b7bc43d1","Type":"ContainerDied","Data":"e787fd4b999aade029525ca31150f0ca69d81e74d922b10588eea746b41a8001"} Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.013419 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vps6w" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.060917 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-dmgvb" podStartSLOduration=3.060895109 podStartE2EDuration="3.060895109s" podCreationTimestamp="2025-11-22 08:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:23.03827038 +0000 UTC m=+1224.642608985" watchObservedRunningTime="2025-11-22 08:23:23.060895109 +0000 UTC m=+1224.665233714" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.165024 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vps6w"] Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.171551 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vps6w"] Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.307168 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05727804-ca8c-48b6-adaf-e86a36263c3c" path="/var/lib/kubelet/pods/05727804-ca8c-48b6-adaf-e86a36263c3c/volumes" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.493224 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.496655 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t22bh\" (UniqueName: \"kubernetes.io/projected/6e8014b5-baf9-4531-a275-2df3ba0c2af1-kube-api-access-t22bh\") pod \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\" (UID: \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\") " Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.496721 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e8014b5-baf9-4531-a275-2df3ba0c2af1-operator-scripts\") pod \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\" (UID: \"6e8014b5-baf9-4531-a275-2df3ba0c2af1\") " Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.497647 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e8014b5-baf9-4531-a275-2df3ba0c2af1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6e8014b5-baf9-4531-a275-2df3ba0c2af1" (UID: "6e8014b5-baf9-4531-a275-2df3ba0c2af1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.505289 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e8014b5-baf9-4531-a275-2df3ba0c2af1-kube-api-access-t22bh" (OuterVolumeSpecName: "kube-api-access-t22bh") pod "6e8014b5-baf9-4531-a275-2df3ba0c2af1" (UID: "6e8014b5-baf9-4531-a275-2df3ba0c2af1"). InnerVolumeSpecName "kube-api-access-t22bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.599870 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t22bh\" (UniqueName: \"kubernetes.io/projected/6e8014b5-baf9-4531-a275-2df3ba0c2af1-kube-api-access-t22bh\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.599901 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e8014b5-baf9-4531-a275-2df3ba0c2af1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.721562 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.728166 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.738701 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.803873 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmjmd\" (UniqueName: \"kubernetes.io/projected/77979ac0-cbb3-4785-9bd5-14cfb29749f3-kube-api-access-pmjmd\") pod \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\" (UID: \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\") " Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.804000 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98303611-37f9-4fee-89fb-c4b58d51ff2f-operator-scripts\") pod \"98303611-37f9-4fee-89fb-c4b58d51ff2f\" (UID: \"98303611-37f9-4fee-89fb-c4b58d51ff2f\") " Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.804061 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf84p\" (UniqueName: \"kubernetes.io/projected/4e7957d0-44ed-4753-a100-253797cb1be9-kube-api-access-gf84p\") pod \"4e7957d0-44ed-4753-a100-253797cb1be9\" (UID: \"4e7957d0-44ed-4753-a100-253797cb1be9\") " Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.804105 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7957d0-44ed-4753-a100-253797cb1be9-operator-scripts\") pod \"4e7957d0-44ed-4753-a100-253797cb1be9\" (UID: \"4e7957d0-44ed-4753-a100-253797cb1be9\") " Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.804139 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77979ac0-cbb3-4785-9bd5-14cfb29749f3-operator-scripts\") pod \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\" (UID: \"77979ac0-cbb3-4785-9bd5-14cfb29749f3\") " Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.804365 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg2wb\" (UniqueName: \"kubernetes.io/projected/98303611-37f9-4fee-89fb-c4b58d51ff2f-kube-api-access-rg2wb\") pod \"98303611-37f9-4fee-89fb-c4b58d51ff2f\" (UID: \"98303611-37f9-4fee-89fb-c4b58d51ff2f\") " Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.804800 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98303611-37f9-4fee-89fb-c4b58d51ff2f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98303611-37f9-4fee-89fb-c4b58d51ff2f" (UID: "98303611-37f9-4fee-89fb-c4b58d51ff2f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.804803 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e7957d0-44ed-4753-a100-253797cb1be9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4e7957d0-44ed-4753-a100-253797cb1be9" (UID: "4e7957d0-44ed-4753-a100-253797cb1be9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.804860 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77979ac0-cbb3-4785-9bd5-14cfb29749f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77979ac0-cbb3-4785-9bd5-14cfb29749f3" (UID: "77979ac0-cbb3-4785-9bd5-14cfb29749f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.805711 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98303611-37f9-4fee-89fb-c4b58d51ff2f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.805816 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7957d0-44ed-4753-a100-253797cb1be9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.805888 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77979ac0-cbb3-4785-9bd5-14cfb29749f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.808801 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77979ac0-cbb3-4785-9bd5-14cfb29749f3-kube-api-access-pmjmd" (OuterVolumeSpecName: "kube-api-access-pmjmd") pod "77979ac0-cbb3-4785-9bd5-14cfb29749f3" (UID: "77979ac0-cbb3-4785-9bd5-14cfb29749f3"). InnerVolumeSpecName "kube-api-access-pmjmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.809434 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e7957d0-44ed-4753-a100-253797cb1be9-kube-api-access-gf84p" (OuterVolumeSpecName: "kube-api-access-gf84p") pod "4e7957d0-44ed-4753-a100-253797cb1be9" (UID: "4e7957d0-44ed-4753-a100-253797cb1be9"). InnerVolumeSpecName "kube-api-access-gf84p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.810611 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98303611-37f9-4fee-89fb-c4b58d51ff2f-kube-api-access-rg2wb" (OuterVolumeSpecName: "kube-api-access-rg2wb") pod "98303611-37f9-4fee-89fb-c4b58d51ff2f" (UID: "98303611-37f9-4fee-89fb-c4b58d51ff2f"). InnerVolumeSpecName "kube-api-access-rg2wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.908371 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmjmd\" (UniqueName: \"kubernetes.io/projected/77979ac0-cbb3-4785-9bd5-14cfb29749f3-kube-api-access-pmjmd\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.908419 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf84p\" (UniqueName: \"kubernetes.io/projected/4e7957d0-44ed-4753-a100-253797cb1be9-kube-api-access-gf84p\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:23 crc kubenswrapper[4735]: I1122 08:23:23.908430 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg2wb\" (UniqueName: \"kubernetes.io/projected/98303611-37f9-4fee-89fb-c4b58d51ff2f-kube-api-access-rg2wb\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.020945 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.021198 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-2f1c-account-create-f4v9s" event={"ID":"4e7957d0-44ed-4753-a100-253797cb1be9","Type":"ContainerDied","Data":"052a7dac6e92065c41f8b425939543dabad99a972b44791188b63338b37da2aa"} Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.021227 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="052a7dac6e92065c41f8b425939543dabad99a972b44791188b63338b37da2aa" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.023813 4735 generic.go:334] "Generic (PLEG): container finished" podID="3f4ab0e4-1b2f-4c1a-8414-d85c38509a64" containerID="406a5f6bf3dcb697458cd3bd8db577b52baa9653d96a6c8c308d44f0a457e919" exitCode=0 Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.023887 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dmgvb" event={"ID":"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64","Type":"ContainerDied","Data":"406a5f6bf3dcb697458cd3bd8db577b52baa9653d96a6c8c308d44f0a457e919"} Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.027205 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.027301 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-nfpfq" event={"ID":"6e8014b5-baf9-4531-a275-2df3ba0c2af1","Type":"ContainerDied","Data":"99a603e2a85a21d64e6a8e6da47115d7139ea6a869994e6b0f162f0fe8d4d80d"} Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.027335 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99a603e2a85a21d64e6a8e6da47115d7139ea6a869994e6b0f162f0fe8d4d80d" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.030647 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c517-account-create-gg4qg" event={"ID":"77979ac0-cbb3-4785-9bd5-14cfb29749f3","Type":"ContainerDied","Data":"a84dc53b42e22c5c5de7c87e82d1e80035b89850d69a872ef257b860f1bbab6a"} Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.030682 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a84dc53b42e22c5c5de7c87e82d1e80035b89850d69a872ef257b860f1bbab6a" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.030745 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c517-account-create-gg4qg" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.042173 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ztsfq" event={"ID":"98303611-37f9-4fee-89fb-c4b58d51ff2f","Type":"ContainerDied","Data":"c5a4480f978fee54aead122316c043b3a0a69ad365c49443c0e055fa67dacfcf"} Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.042232 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5a4480f978fee54aead122316c043b3a0a69ad365c49443c0e055fa67dacfcf" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.042382 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ztsfq" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.049537 4735 generic.go:334] "Generic (PLEG): container finished" podID="c5e4fc11-5ae0-4336-bc26-f07ca5230b48" containerID="a5bcee5f440f8e179c2ad85fc45661bf74347ae9960f818494e1ba614685722d" exitCode=0 Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.051828 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-20d8-account-create-crjxr" event={"ID":"c5e4fc11-5ae0-4336-bc26-f07ca5230b48","Type":"ContainerDied","Data":"a5bcee5f440f8e179c2ad85fc45661bf74347ae9960f818494e1ba614685722d"} Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.481514 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.526556 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/659efe09-eb81-4668-a68f-fc72b7bc43d1-operator-scripts\") pod \"659efe09-eb81-4668-a68f-fc72b7bc43d1\" (UID: \"659efe09-eb81-4668-a68f-fc72b7bc43d1\") " Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.526799 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tlpv\" (UniqueName: \"kubernetes.io/projected/659efe09-eb81-4668-a68f-fc72b7bc43d1-kube-api-access-6tlpv\") pod \"659efe09-eb81-4668-a68f-fc72b7bc43d1\" (UID: \"659efe09-eb81-4668-a68f-fc72b7bc43d1\") " Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.531139 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/659efe09-eb81-4668-a68f-fc72b7bc43d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "659efe09-eb81-4668-a68f-fc72b7bc43d1" (UID: "659efe09-eb81-4668-a68f-fc72b7bc43d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.537815 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/659efe09-eb81-4668-a68f-fc72b7bc43d1-kube-api-access-6tlpv" (OuterVolumeSpecName: "kube-api-access-6tlpv") pod "659efe09-eb81-4668-a68f-fc72b7bc43d1" (UID: "659efe09-eb81-4668-a68f-fc72b7bc43d1"). InnerVolumeSpecName "kube-api-access-6tlpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.542570 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.629171 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z287l\" (UniqueName: \"kubernetes.io/projected/cf0ea764-b639-4959-8a27-e2f674be3329-kube-api-access-z287l\") pod \"cf0ea764-b639-4959-8a27-e2f674be3329\" (UID: \"cf0ea764-b639-4959-8a27-e2f674be3329\") " Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.629258 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0ea764-b639-4959-8a27-e2f674be3329-operator-scripts\") pod \"cf0ea764-b639-4959-8a27-e2f674be3329\" (UID: \"cf0ea764-b639-4959-8a27-e2f674be3329\") " Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.629674 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/659efe09-eb81-4668-a68f-fc72b7bc43d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.629701 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tlpv\" (UniqueName: \"kubernetes.io/projected/659efe09-eb81-4668-a68f-fc72b7bc43d1-kube-api-access-6tlpv\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.630091 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf0ea764-b639-4959-8a27-e2f674be3329-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf0ea764-b639-4959-8a27-e2f674be3329" (UID: "cf0ea764-b639-4959-8a27-e2f674be3329"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.633698 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf0ea764-b639-4959-8a27-e2f674be3329-kube-api-access-z287l" (OuterVolumeSpecName: "kube-api-access-z287l") pod "cf0ea764-b639-4959-8a27-e2f674be3329" (UID: "cf0ea764-b639-4959-8a27-e2f674be3329"). InnerVolumeSpecName "kube-api-access-z287l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.731386 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z287l\" (UniqueName: \"kubernetes.io/projected/cf0ea764-b639-4959-8a27-e2f674be3329-kube-api-access-z287l\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:24 crc kubenswrapper[4735]: I1122 08:23:24.731425 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0ea764-b639-4959-8a27-e2f674be3329-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:25 crc kubenswrapper[4735]: I1122 08:23:25.066284 4735 generic.go:334] "Generic (PLEG): container finished" podID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerID="639b9585adb4796e0e70cf58fd13241ab15ccbd2b150f54e3680a54b17b7030f" exitCode=0 Nov 22 08:23:25 crc kubenswrapper[4735]: I1122 08:23:25.066363 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10","Type":"ContainerDied","Data":"639b9585adb4796e0e70cf58fd13241ab15ccbd2b150f54e3680a54b17b7030f"} Nov 22 08:23:25 crc kubenswrapper[4735]: I1122 08:23:25.068130 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2n7mj" event={"ID":"cf0ea764-b639-4959-8a27-e2f674be3329","Type":"ContainerDied","Data":"9ced673df6504baf62400eb1bf5fbe637e3f7623781f4dfeca97a4459e0f98d4"} Nov 22 08:23:25 crc kubenswrapper[4735]: I1122 08:23:25.068164 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ced673df6504baf62400eb1bf5fbe637e3f7623781f4dfeca97a4459e0f98d4" Nov 22 08:23:25 crc kubenswrapper[4735]: I1122 08:23:25.068216 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2n7mj" Nov 22 08:23:25 crc kubenswrapper[4735]: I1122 08:23:25.085723 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f25a-account-create-rg9dx" Nov 22 08:23:25 crc kubenswrapper[4735]: I1122 08:23:25.087550 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f25a-account-create-rg9dx" event={"ID":"659efe09-eb81-4668-a68f-fc72b7bc43d1","Type":"ContainerDied","Data":"614b89b3383419337c9604506293dc47428b789bd63bea663eb10431193e5966"} Nov 22 08:23:25 crc kubenswrapper[4735]: I1122 08:23:25.087662 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="614b89b3383419337c9604506293dc47428b789bd63bea663eb10431193e5966" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.095782 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-20d8-account-create-crjxr" event={"ID":"c5e4fc11-5ae0-4336-bc26-f07ca5230b48","Type":"ContainerDied","Data":"8c4a8dc3c2d5ed0c190d8202f28f065db34ab962c34ba47bb3832daaeb248098"} Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.096098 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c4a8dc3c2d5ed0c190d8202f28f065db34ab962c34ba47bb3832daaeb248098" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.117377 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dmgvb" event={"ID":"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64","Type":"ContainerDied","Data":"8414ecf2510e6deecb481777cb0b87f65cdd219281e18d3480095c0221bc8d82"} Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.117434 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8414ecf2510e6deecb481777cb0b87f65cdd219281e18d3480095c0221bc8d82" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.192988 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.297352 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.377201 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqt8w\" (UniqueName: \"kubernetes.io/projected/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-kube-api-access-qqt8w\") pod \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\" (UID: \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\") " Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.377249 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-operator-scripts\") pod \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\" (UID: \"3f4ab0e4-1b2f-4c1a-8414-d85c38509a64\") " Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.377875 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f4ab0e4-1b2f-4c1a-8414-d85c38509a64" (UID: "3f4ab0e4-1b2f-4c1a-8414-d85c38509a64"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.389633 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-kube-api-access-qqt8w" (OuterVolumeSpecName: "kube-api-access-qqt8w") pod "3f4ab0e4-1b2f-4c1a-8414-d85c38509a64" (UID: "3f4ab0e4-1b2f-4c1a-8414-d85c38509a64"). InnerVolumeSpecName "kube-api-access-qqt8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.479678 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-operator-scripts\") pod \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\" (UID: \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\") " Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.479959 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzckl\" (UniqueName: \"kubernetes.io/projected/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-kube-api-access-zzckl\") pod \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\" (UID: \"c5e4fc11-5ae0-4336-bc26-f07ca5230b48\") " Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.480221 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5e4fc11-5ae0-4336-bc26-f07ca5230b48" (UID: "c5e4fc11-5ae0-4336-bc26-f07ca5230b48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.480683 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqt8w\" (UniqueName: \"kubernetes.io/projected/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-kube-api-access-qqt8w\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.480707 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.480722 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.484508 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-kube-api-access-zzckl" (OuterVolumeSpecName: "kube-api-access-zzckl") pod "c5e4fc11-5ae0-4336-bc26-f07ca5230b48" (UID: "c5e4fc11-5ae0-4336-bc26-f07ca5230b48"). InnerVolumeSpecName "kube-api-access-zzckl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.551869 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-m9x5r"] Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552303 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf0ea764-b639-4959-8a27-e2f674be3329" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552320 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf0ea764-b639-4959-8a27-e2f674be3329" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552333 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77979ac0-cbb3-4785-9bd5-14cfb29749f3" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552340 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="77979ac0-cbb3-4785-9bd5-14cfb29749f3" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552354 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5e4fc11-5ae0-4336-bc26-f07ca5230b48" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552362 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5e4fc11-5ae0-4336-bc26-f07ca5230b48" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552375 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e8014b5-baf9-4531-a275-2df3ba0c2af1" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552380 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e8014b5-baf9-4531-a275-2df3ba0c2af1" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552388 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e7957d0-44ed-4753-a100-253797cb1be9" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552394 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e7957d0-44ed-4753-a100-253797cb1be9" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552405 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05727804-ca8c-48b6-adaf-e86a36263c3c" containerName="dnsmasq-dns" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552411 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="05727804-ca8c-48b6-adaf-e86a36263c3c" containerName="dnsmasq-dns" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552417 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98303611-37f9-4fee-89fb-c4b58d51ff2f" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552424 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="98303611-37f9-4fee-89fb-c4b58d51ff2f" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552444 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659efe09-eb81-4668-a68f-fc72b7bc43d1" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552450 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="659efe09-eb81-4668-a68f-fc72b7bc43d1" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552478 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4ab0e4-1b2f-4c1a-8414-d85c38509a64" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552486 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4ab0e4-1b2f-4c1a-8414-d85c38509a64" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: E1122 08:23:26.552500 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05727804-ca8c-48b6-adaf-e86a36263c3c" containerName="init" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552506 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="05727804-ca8c-48b6-adaf-e86a36263c3c" containerName="init" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552699 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf0ea764-b639-4959-8a27-e2f674be3329" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552714 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e8014b5-baf9-4531-a275-2df3ba0c2af1" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552723 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="05727804-ca8c-48b6-adaf-e86a36263c3c" containerName="dnsmasq-dns" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552737 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="98303611-37f9-4fee-89fb-c4b58d51ff2f" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552749 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5e4fc11-5ae0-4336-bc26-f07ca5230b48" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552764 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="77979ac0-cbb3-4785-9bd5-14cfb29749f3" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552771 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="659efe09-eb81-4668-a68f-fc72b7bc43d1" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552783 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e7957d0-44ed-4753-a100-253797cb1be9" containerName="mariadb-account-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.552796 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f4ab0e4-1b2f-4c1a-8414-d85c38509a64" containerName="mariadb-database-create" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.553428 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.555599 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n5fwv" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.567952 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.572814 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-m9x5r"] Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.582781 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-combined-ca-bundle\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.582900 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-db-sync-config-data\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.582952 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-config-data\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.583234 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twfqg\" (UniqueName: \"kubernetes.io/projected/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-kube-api-access-twfqg\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.583449 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzckl\" (UniqueName: \"kubernetes.io/projected/c5e4fc11-5ae0-4336-bc26-f07ca5230b48-kube-api-access-zzckl\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.684352 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twfqg\" (UniqueName: \"kubernetes.io/projected/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-kube-api-access-twfqg\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.684502 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-combined-ca-bundle\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.684546 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-db-sync-config-data\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.684569 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-config-data\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.688574 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-db-sync-config-data\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.688647 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-combined-ca-bundle\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.688813 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-config-data\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.708557 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twfqg\" (UniqueName: \"kubernetes.io/projected/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-kube-api-access-twfqg\") pod \"glance-db-sync-m9x5r\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:26 crc kubenswrapper[4735]: I1122 08:23:26.869259 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-m9x5r" Nov 22 08:23:27 crc kubenswrapper[4735]: I1122 08:23:27.133187 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10","Type":"ContainerStarted","Data":"c844c326444f1cb2205eb64a4c985689b5a47f324d64e8e485fb7134dbd49654"} Nov 22 08:23:27 crc kubenswrapper[4735]: I1122 08:23:27.133662 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 22 08:23:27 crc kubenswrapper[4735]: I1122 08:23:27.137529 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-20d8-account-create-crjxr" Nov 22 08:23:27 crc kubenswrapper[4735]: I1122 08:23:27.138555 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerStarted","Data":"37cc336eeb0e55dacf3617c3e737ce28e02bacb73aafccee97cb3bc4b4e0c6d2"} Nov 22 08:23:27 crc kubenswrapper[4735]: I1122 08:23:27.138661 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dmgvb" Nov 22 08:23:27 crc kubenswrapper[4735]: I1122 08:23:27.163020 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.044719617 podStartE2EDuration="1m1.163005651s" podCreationTimestamp="2025-11-22 08:22:26 +0000 UTC" firstStartedPulling="2025-11-22 08:22:28.108923173 +0000 UTC m=+1169.713261778" lastFinishedPulling="2025-11-22 08:22:50.227209157 +0000 UTC m=+1191.831547812" observedRunningTime="2025-11-22 08:23:27.15541231 +0000 UTC m=+1228.759750925" watchObservedRunningTime="2025-11-22 08:23:27.163005651 +0000 UTC m=+1228.767344256" Nov 22 08:23:27 crc kubenswrapper[4735]: I1122 08:23:27.192056 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=19.324611532 podStartE2EDuration="54.192038908s" podCreationTimestamp="2025-11-22 08:22:33 +0000 UTC" firstStartedPulling="2025-11-22 08:22:51.298385309 +0000 UTC m=+1192.902723914" lastFinishedPulling="2025-11-22 08:23:26.165812685 +0000 UTC m=+1227.770151290" observedRunningTime="2025-11-22 08:23:27.187269626 +0000 UTC m=+1228.791608241" watchObservedRunningTime="2025-11-22 08:23:27.192038908 +0000 UTC m=+1228.796377513" Nov 22 08:23:27 crc kubenswrapper[4735]: I1122 08:23:27.482860 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-m9x5r"] Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.151408 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-m9x5r" event={"ID":"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3","Type":"ContainerStarted","Data":"301efc5175d965eebd6c21df22788f260d29a3a03feb01be64d3974a9c0ff3a1"} Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.208137 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-8drgk"] Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.209529 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.222906 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-8drgk"] Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.323175 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4drqd\" (UniqueName: \"kubernetes.io/projected/c6d00bef-4c5e-40f0-9c2f-8811da61708c-kube-api-access-4drqd\") pod \"mysqld-exporter-openstack-cell1-db-create-8drgk\" (UID: \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.323603 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6d00bef-4c5e-40f0-9c2f-8811da61708c-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-8drgk\" (UID: \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.425550 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6d00bef-4c5e-40f0-9c2f-8811da61708c-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-8drgk\" (UID: \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.425631 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4drqd\" (UniqueName: \"kubernetes.io/projected/c6d00bef-4c5e-40f0-9c2f-8811da61708c-kube-api-access-4drqd\") pod \"mysqld-exporter-openstack-cell1-db-create-8drgk\" (UID: \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.426991 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6d00bef-4c5e-40f0-9c2f-8811da61708c-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-8drgk\" (UID: \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.440921 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-85db-account-create-rt6xj"] Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.442307 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.447078 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.455147 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4drqd\" (UniqueName: \"kubernetes.io/projected/c6d00bef-4c5e-40f0-9c2f-8811da61708c-kube-api-access-4drqd\") pod \"mysqld-exporter-openstack-cell1-db-create-8drgk\" (UID: \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.462058 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-85db-account-create-rt6xj"] Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.527202 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a31925-48f6-4fba-91e0-b5884bde3344-operator-scripts\") pod \"mysqld-exporter-85db-account-create-rt6xj\" (UID: \"30a31925-48f6-4fba-91e0-b5884bde3344\") " pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.527291 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgjbk\" (UniqueName: \"kubernetes.io/projected/30a31925-48f6-4fba-91e0-b5884bde3344-kube-api-access-cgjbk\") pod \"mysqld-exporter-85db-account-create-rt6xj\" (UID: \"30a31925-48f6-4fba-91e0-b5884bde3344\") " pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.528317 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.629579 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a31925-48f6-4fba-91e0-b5884bde3344-operator-scripts\") pod \"mysqld-exporter-85db-account-create-rt6xj\" (UID: \"30a31925-48f6-4fba-91e0-b5884bde3344\") " pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.629685 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgjbk\" (UniqueName: \"kubernetes.io/projected/30a31925-48f6-4fba-91e0-b5884bde3344-kube-api-access-cgjbk\") pod \"mysqld-exporter-85db-account-create-rt6xj\" (UID: \"30a31925-48f6-4fba-91e0-b5884bde3344\") " pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.631006 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a31925-48f6-4fba-91e0-b5884bde3344-operator-scripts\") pod \"mysqld-exporter-85db-account-create-rt6xj\" (UID: \"30a31925-48f6-4fba-91e0-b5884bde3344\") " pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.648377 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgjbk\" (UniqueName: \"kubernetes.io/projected/30a31925-48f6-4fba-91e0-b5884bde3344-kube-api-access-cgjbk\") pod \"mysqld-exporter-85db-account-create-rt6xj\" (UID: \"30a31925-48f6-4fba-91e0-b5884bde3344\") " pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:28 crc kubenswrapper[4735]: I1122 08:23:28.831856 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:29 crc kubenswrapper[4735]: I1122 08:23:29.091736 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-8drgk"] Nov 22 08:23:29 crc kubenswrapper[4735]: I1122 08:23:29.172056 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" event={"ID":"c6d00bef-4c5e-40f0-9c2f-8811da61708c","Type":"ContainerStarted","Data":"f99b10cadb72e2a8f4b6e6ef17c5b0c625fa5fa2413ec1efeea99966bb875e9b"} Nov 22 08:23:29 crc kubenswrapper[4735]: I1122 08:23:29.178169 4735 generic.go:334] "Generic (PLEG): container finished" podID="abf32873-0945-47b6-8f4f-ed1b18dc6eb1" containerID="0ff683dd6063093793ecd3f43038e6e3e4b89d6469aaaefc88bf285ffadc0a6a" exitCode=0 Nov 22 08:23:29 crc kubenswrapper[4735]: I1122 08:23:29.178210 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hvj22" event={"ID":"abf32873-0945-47b6-8f4f-ed1b18dc6eb1","Type":"ContainerDied","Data":"0ff683dd6063093793ecd3f43038e6e3e4b89d6469aaaefc88bf285ffadc0a6a"} Nov 22 08:23:29 crc kubenswrapper[4735]: I1122 08:23:29.334372 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-85db-account-create-rt6xj"] Nov 22 08:23:29 crc kubenswrapper[4735]: W1122 08:23:29.334597 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30a31925_48f6_4fba_91e0_b5884bde3344.slice/crio-a67d7179e5ba7ae0a3d8bf0b4dea692e01e71761413088b840cbac8811a96567 WatchSource:0}: Error finding container a67d7179e5ba7ae0a3d8bf0b4dea692e01e71761413088b840cbac8811a96567: Status 404 returned error can't find the container with id a67d7179e5ba7ae0a3d8bf0b4dea692e01e71761413088b840cbac8811a96567 Nov 22 08:23:29 crc kubenswrapper[4735]: I1122 08:23:29.533566 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.188678 4735 generic.go:334] "Generic (PLEG): container finished" podID="c6d00bef-4c5e-40f0-9c2f-8811da61708c" containerID="fc3c3652534f48f5bfe73cefe808e25bf3d0c43f4d74ffdb893b753719480640" exitCode=0 Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.189113 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" event={"ID":"c6d00bef-4c5e-40f0-9c2f-8811da61708c","Type":"ContainerDied","Data":"fc3c3652534f48f5bfe73cefe808e25bf3d0c43f4d74ffdb893b753719480640"} Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.190847 4735 generic.go:334] "Generic (PLEG): container finished" podID="30a31925-48f6-4fba-91e0-b5884bde3344" containerID="f85ac7ec786dbd469571c9691a06c4c6bc7cda82287dd0591119f0c7d1c3510c" exitCode=0 Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.191084 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-85db-account-create-rt6xj" event={"ID":"30a31925-48f6-4fba-91e0-b5884bde3344","Type":"ContainerDied","Data":"f85ac7ec786dbd469571c9691a06c4c6bc7cda82287dd0591119f0c7d1c3510c"} Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.191106 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-85db-account-create-rt6xj" event={"ID":"30a31925-48f6-4fba-91e0-b5884bde3344","Type":"ContainerStarted","Data":"a67d7179e5ba7ae0a3d8bf0b4dea692e01e71761413088b840cbac8811a96567"} Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.653265 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.661750 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.780543 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-ring-data-devices\") pod \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.780627 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-scripts\") pod \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.780704 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-etc-swift\") pod \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.780804 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-dispersionconf\") pod \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.780857 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-swiftconf\") pod \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.780910 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxfbr\" (UniqueName: \"kubernetes.io/projected/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-kube-api-access-xxfbr\") pod \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.780982 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-combined-ca-bundle\") pod \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\" (UID: \"abf32873-0945-47b6-8f4f-ed1b18dc6eb1\") " Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.782155 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "abf32873-0945-47b6-8f4f-ed1b18dc6eb1" (UID: "abf32873-0945-47b6-8f4f-ed1b18dc6eb1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.783202 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "abf32873-0945-47b6-8f4f-ed1b18dc6eb1" (UID: "abf32873-0945-47b6-8f4f-ed1b18dc6eb1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.789861 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-kube-api-access-xxfbr" (OuterVolumeSpecName: "kube-api-access-xxfbr") pod "abf32873-0945-47b6-8f4f-ed1b18dc6eb1" (UID: "abf32873-0945-47b6-8f4f-ed1b18dc6eb1"). InnerVolumeSpecName "kube-api-access-xxfbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.792526 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "abf32873-0945-47b6-8f4f-ed1b18dc6eb1" (UID: "abf32873-0945-47b6-8f4f-ed1b18dc6eb1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.814647 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "abf32873-0945-47b6-8f4f-ed1b18dc6eb1" (UID: "abf32873-0945-47b6-8f4f-ed1b18dc6eb1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.818160 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-scripts" (OuterVolumeSpecName: "scripts") pod "abf32873-0945-47b6-8f4f-ed1b18dc6eb1" (UID: "abf32873-0945-47b6-8f4f-ed1b18dc6eb1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.825957 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abf32873-0945-47b6-8f4f-ed1b18dc6eb1" (UID: "abf32873-0945-47b6-8f4f-ed1b18dc6eb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.883863 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.883909 4735 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.883921 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.883934 4735 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.883945 4735 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.883956 4735 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:30 crc kubenswrapper[4735]: I1122 08:23:30.883968 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxfbr\" (UniqueName: \"kubernetes.io/projected/abf32873-0945-47b6-8f4f-ed1b18dc6eb1-kube-api-access-xxfbr\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.211125 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hvj22" Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.211273 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hvj22" event={"ID":"abf32873-0945-47b6-8f4f-ed1b18dc6eb1","Type":"ContainerDied","Data":"3a89ceb6f55c9e6e788f9307665f29ca82a875ad03c7abbe4b70b81cd7ac2681"} Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.211642 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a89ceb6f55c9e6e788f9307665f29ca82a875ad03c7abbe4b70b81cd7ac2681" Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.371892 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z5csg" podUID="d99e6f50-a748-4ae8-b1b0-9df901b37201" containerName="ovn-controller" probeResult="failure" output=< Nov 22 08:23:31 crc kubenswrapper[4735]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 22 08:23:31 crc kubenswrapper[4735]: > Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.840355 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.851065 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.908769 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgjbk\" (UniqueName: \"kubernetes.io/projected/30a31925-48f6-4fba-91e0-b5884bde3344-kube-api-access-cgjbk\") pod \"30a31925-48f6-4fba-91e0-b5884bde3344\" (UID: \"30a31925-48f6-4fba-91e0-b5884bde3344\") " Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.909340 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a31925-48f6-4fba-91e0-b5884bde3344-operator-scripts\") pod \"30a31925-48f6-4fba-91e0-b5884bde3344\" (UID: \"30a31925-48f6-4fba-91e0-b5884bde3344\") " Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.909938 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a31925-48f6-4fba-91e0-b5884bde3344-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30a31925-48f6-4fba-91e0-b5884bde3344" (UID: "30a31925-48f6-4fba-91e0-b5884bde3344"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.910664 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a31925-48f6-4fba-91e0-b5884bde3344-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:31 crc kubenswrapper[4735]: I1122 08:23:31.913643 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a31925-48f6-4fba-91e0-b5884bde3344-kube-api-access-cgjbk" (OuterVolumeSpecName: "kube-api-access-cgjbk") pod "30a31925-48f6-4fba-91e0-b5884bde3344" (UID: "30a31925-48f6-4fba-91e0-b5884bde3344"). InnerVolumeSpecName "kube-api-access-cgjbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.012152 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4drqd\" (UniqueName: \"kubernetes.io/projected/c6d00bef-4c5e-40f0-9c2f-8811da61708c-kube-api-access-4drqd\") pod \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\" (UID: \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\") " Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.012248 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6d00bef-4c5e-40f0-9c2f-8811da61708c-operator-scripts\") pod \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\" (UID: \"c6d00bef-4c5e-40f0-9c2f-8811da61708c\") " Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.012732 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6d00bef-4c5e-40f0-9c2f-8811da61708c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c6d00bef-4c5e-40f0-9c2f-8811da61708c" (UID: "c6d00bef-4c5e-40f0-9c2f-8811da61708c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.013179 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgjbk\" (UniqueName: \"kubernetes.io/projected/30a31925-48f6-4fba-91e0-b5884bde3344-kube-api-access-cgjbk\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.013202 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6d00bef-4c5e-40f0-9c2f-8811da61708c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.027118 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6d00bef-4c5e-40f0-9c2f-8811da61708c-kube-api-access-4drqd" (OuterVolumeSpecName: "kube-api-access-4drqd") pod "c6d00bef-4c5e-40f0-9c2f-8811da61708c" (UID: "c6d00bef-4c5e-40f0-9c2f-8811da61708c"). InnerVolumeSpecName "kube-api-access-4drqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.114566 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4drqd\" (UniqueName: \"kubernetes.io/projected/c6d00bef-4c5e-40f0-9c2f-8811da61708c-kube-api-access-4drqd\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.245739 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-85db-account-create-rt6xj" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.245747 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-85db-account-create-rt6xj" event={"ID":"30a31925-48f6-4fba-91e0-b5884bde3344","Type":"ContainerDied","Data":"a67d7179e5ba7ae0a3d8bf0b4dea692e01e71761413088b840cbac8811a96567"} Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.245876 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a67d7179e5ba7ae0a3d8bf0b4dea692e01e71761413088b840cbac8811a96567" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.247677 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" event={"ID":"c6d00bef-4c5e-40f0-9c2f-8811da61708c","Type":"ContainerDied","Data":"f99b10cadb72e2a8f4b6e6ef17c5b0c625fa5fa2413ec1efeea99966bb875e9b"} Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.247715 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f99b10cadb72e2a8f4b6e6ef17c5b0c625fa5fa2413ec1efeea99966bb875e9b" Nov 22 08:23:32 crc kubenswrapper[4735]: I1122 08:23:32.247761 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-8drgk" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.665375 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:23:33 crc kubenswrapper[4735]: E1122 08:23:33.666175 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6d00bef-4c5e-40f0-9c2f-8811da61708c" containerName="mariadb-database-create" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.666192 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6d00bef-4c5e-40f0-9c2f-8811da61708c" containerName="mariadb-database-create" Nov 22 08:23:33 crc kubenswrapper[4735]: E1122 08:23:33.666212 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a31925-48f6-4fba-91e0-b5884bde3344" containerName="mariadb-account-create" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.666219 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a31925-48f6-4fba-91e0-b5884bde3344" containerName="mariadb-account-create" Nov 22 08:23:33 crc kubenswrapper[4735]: E1122 08:23:33.666262 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf32873-0945-47b6-8f4f-ed1b18dc6eb1" containerName="swift-ring-rebalance" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.666270 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf32873-0945-47b6-8f4f-ed1b18dc6eb1" containerName="swift-ring-rebalance" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.666514 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf32873-0945-47b6-8f4f-ed1b18dc6eb1" containerName="swift-ring-rebalance" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.666534 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6d00bef-4c5e-40f0-9c2f-8811da61708c" containerName="mariadb-database-create" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.666549 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a31925-48f6-4fba-91e0-b5884bde3344" containerName="mariadb-account-create" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.667365 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.670399 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.683063 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.743945 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-config-data\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.744095 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f8sz\" (UniqueName: \"kubernetes.io/projected/3b29201e-860c-4944-ad29-06abb7aed0c6-kube-api-access-6f8sz\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.744277 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.846621 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.846686 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-config-data\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.846777 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f8sz\" (UniqueName: \"kubernetes.io/projected/3b29201e-860c-4944-ad29-06abb7aed0c6-kube-api-access-6f8sz\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.859422 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-config-data\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.865439 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.866476 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f8sz\" (UniqueName: \"kubernetes.io/projected/3b29201e-860c-4944-ad29-06abb7aed0c6-kube-api-access-6f8sz\") pod \"mysqld-exporter-0\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " pod="openstack/mysqld-exporter-0" Nov 22 08:23:33 crc kubenswrapper[4735]: I1122 08:23:33.996925 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 22 08:23:34 crc kubenswrapper[4735]: I1122 08:23:34.534585 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:34 crc kubenswrapper[4735]: I1122 08:23:34.538336 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:34 crc kubenswrapper[4735]: I1122 08:23:34.599925 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:23:34 crc kubenswrapper[4735]: W1122 08:23:34.601922 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b29201e_860c_4944_ad29_06abb7aed0c6.slice/crio-4da5c46aa8080594d1437e5b10e078828e7b5a8c8bedb8c0c250d3eeb35cbedd WatchSource:0}: Error finding container 4da5c46aa8080594d1437e5b10e078828e7b5a8c8bedb8c0c250d3eeb35cbedd: Status 404 returned error can't find the container with id 4da5c46aa8080594d1437e5b10e078828e7b5a8c8bedb8c0c250d3eeb35cbedd Nov 22 08:23:35 crc kubenswrapper[4735]: I1122 08:23:35.293835 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"3b29201e-860c-4944-ad29-06abb7aed0c6","Type":"ContainerStarted","Data":"4da5c46aa8080594d1437e5b10e078828e7b5a8c8bedb8c0c250d3eeb35cbedd"} Nov 22 08:23:35 crc kubenswrapper[4735]: I1122 08:23:35.295788 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:35 crc kubenswrapper[4735]: I1122 08:23:35.793651 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:35 crc kubenswrapper[4735]: I1122 08:23:35.802853 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49e1ddee-0d04-4d19-9c87-e438e5e70e70-etc-swift\") pod \"swift-storage-0\" (UID: \"49e1ddee-0d04-4d19-9c87-e438e5e70e70\") " pod="openstack/swift-storage-0" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.021586 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.243591 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z5csg" podUID="d99e6f50-a748-4ae8-b1b0-9df901b37201" containerName="ovn-controller" probeResult="failure" output=< Nov 22 08:23:36 crc kubenswrapper[4735]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 22 08:23:36 crc kubenswrapper[4735]: > Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.269005 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.277225 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7mhwj" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.311390 4735 generic.go:334] "Generic (PLEG): container finished" podID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerID="b6b68c78ecac956d2a32b20d00cbfd5998f52fb623cb15919609639792d34e37" exitCode=0 Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.312105 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5ee793a0-c437-47a3-b05d-369a1730d6b1","Type":"ContainerDied","Data":"b6b68c78ecac956d2a32b20d00cbfd5998f52fb623cb15919609639792d34e37"} Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.536550 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z5csg-config-nl7bp"] Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.538018 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.540963 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.556945 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z5csg-config-nl7bp"] Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.616294 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run-ovn\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.616844 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-additional-scripts\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.616891 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-log-ovn\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.616921 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.616970 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9jkq\" (UniqueName: \"kubernetes.io/projected/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-kube-api-access-l9jkq\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.617015 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-scripts\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: W1122 08:23:36.656696 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49e1ddee_0d04_4d19_9c87_e438e5e70e70.slice/crio-70b6d7cc6db2f5d332eff435e67e36f524e1021ab85c6ffdfb49f3b791b23ab5 WatchSource:0}: Error finding container 70b6d7cc6db2f5d332eff435e67e36f524e1021ab85c6ffdfb49f3b791b23ab5: Status 404 returned error can't find the container with id 70b6d7cc6db2f5d332eff435e67e36f524e1021ab85c6ffdfb49f3b791b23ab5 Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.658170 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.718902 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run-ovn\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.719045 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-additional-scripts\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.719077 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-log-ovn\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.719104 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.719149 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9jkq\" (UniqueName: \"kubernetes.io/projected/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-kube-api-access-l9jkq\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.719191 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-scripts\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.719314 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run-ovn\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.719393 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-log-ovn\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.720142 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-additional-scripts\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.720216 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.723191 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-scripts\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.746337 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9jkq\" (UniqueName: \"kubernetes.io/projected/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-kube-api-access-l9jkq\") pod \"ovn-controller-z5csg-config-nl7bp\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:36 crc kubenswrapper[4735]: I1122 08:23:36.858891 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.321445 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"70b6d7cc6db2f5d332eff435e67e36f524e1021ab85c6ffdfb49f3b791b23ab5"} Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.393995 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z5csg-config-nl7bp"] Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.454941 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.827974 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-7cghr"] Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.829302 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.849811 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-7cghr"] Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.933787 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-9cd9-account-create-zxr8b"] Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.935247 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.941753 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebfceb98-899b-40d4-a27d-dfb27042557e-operator-scripts\") pod \"cinder-db-create-7cghr\" (UID: \"ebfceb98-899b-40d4-a27d-dfb27042557e\") " pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.941946 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjxjx\" (UniqueName: \"kubernetes.io/projected/ebfceb98-899b-40d4-a27d-dfb27042557e-kube-api-access-xjxjx\") pod \"cinder-db-create-7cghr\" (UID: \"ebfceb98-899b-40d4-a27d-dfb27042557e\") " pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.945207 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 22 08:23:37 crc kubenswrapper[4735]: I1122 08:23:37.945366 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-9cd9-account-create-zxr8b"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.049263 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-bkgcz"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.051448 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.090017 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-098a-account-create-7sqdp"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.092904 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.094873 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-operator-scripts\") pod \"heat-db-create-bkgcz\" (UID: \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\") " pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.094971 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjxjx\" (UniqueName: \"kubernetes.io/projected/ebfceb98-899b-40d4-a27d-dfb27042557e-kube-api-access-xjxjx\") pod \"cinder-db-create-7cghr\" (UID: \"ebfceb98-899b-40d4-a27d-dfb27042557e\") " pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.095001 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebfceb98-899b-40d4-a27d-dfb27042557e-operator-scripts\") pod \"cinder-db-create-7cghr\" (UID: \"ebfceb98-899b-40d4-a27d-dfb27042557e\") " pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.095260 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-operator-scripts\") pod \"heat-9cd9-account-create-zxr8b\" (UID: \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\") " pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.095312 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf6x2\" (UniqueName: \"kubernetes.io/projected/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-kube-api-access-bf6x2\") pod \"heat-db-create-bkgcz\" (UID: \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\") " pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.095360 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9zxr\" (UniqueName: \"kubernetes.io/projected/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-kube-api-access-x9zxr\") pod \"heat-9cd9-account-create-zxr8b\" (UID: \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\") " pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.096070 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebfceb98-899b-40d4-a27d-dfb27042557e-operator-scripts\") pod \"cinder-db-create-7cghr\" (UID: \"ebfceb98-899b-40d4-a27d-dfb27042557e\") " pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.100696 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.112476 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-bkgcz"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.128090 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-098a-account-create-7sqdp"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.149178 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjxjx\" (UniqueName: \"kubernetes.io/projected/ebfceb98-899b-40d4-a27d-dfb27042557e-kube-api-access-xjxjx\") pod \"cinder-db-create-7cghr\" (UID: \"ebfceb98-899b-40d4-a27d-dfb27042557e\") " pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.199744 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9zxr\" (UniqueName: \"kubernetes.io/projected/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-kube-api-access-x9zxr\") pod \"heat-9cd9-account-create-zxr8b\" (UID: \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\") " pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.200111 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4925fb22-c2bb-4609-a593-5589e9137ecc-operator-scripts\") pod \"cinder-098a-account-create-7sqdp\" (UID: \"4925fb22-c2bb-4609-a593-5589e9137ecc\") " pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.200233 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp9rh\" (UniqueName: \"kubernetes.io/projected/4925fb22-c2bb-4609-a593-5589e9137ecc-kube-api-access-xp9rh\") pod \"cinder-098a-account-create-7sqdp\" (UID: \"4925fb22-c2bb-4609-a593-5589e9137ecc\") " pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.200362 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-operator-scripts\") pod \"heat-db-create-bkgcz\" (UID: \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\") " pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.200547 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-operator-scripts\") pod \"heat-9cd9-account-create-zxr8b\" (UID: \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\") " pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.200650 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf6x2\" (UniqueName: \"kubernetes.io/projected/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-kube-api-access-bf6x2\") pod \"heat-db-create-bkgcz\" (UID: \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\") " pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.201240 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-operator-scripts\") pod \"heat-db-create-bkgcz\" (UID: \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\") " pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.204937 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-operator-scripts\") pod \"heat-9cd9-account-create-zxr8b\" (UID: \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\") " pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.222937 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-9grv6"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.224866 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.230825 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.231075 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.232184 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.232767 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9zxr\" (UniqueName: \"kubernetes.io/projected/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-kube-api-access-x9zxr\") pod \"heat-9cd9-account-create-zxr8b\" (UID: \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\") " pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.233001 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s4xtc" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.235408 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf6x2\" (UniqueName: \"kubernetes.io/projected/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-kube-api-access-bf6x2\") pod \"heat-db-create-bkgcz\" (UID: \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\") " pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.238993 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9grv6"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.249833 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.302542 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-combined-ca-bundle\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.302981 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4925fb22-c2bb-4609-a593-5589e9137ecc-operator-scripts\") pod \"cinder-098a-account-create-7sqdp\" (UID: \"4925fb22-c2bb-4609-a593-5589e9137ecc\") " pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.303030 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp9rh\" (UniqueName: \"kubernetes.io/projected/4925fb22-c2bb-4609-a593-5589e9137ecc-kube-api-access-xp9rh\") pod \"cinder-098a-account-create-7sqdp\" (UID: \"4925fb22-c2bb-4609-a593-5589e9137ecc\") " pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.303070 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-config-data\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.303111 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cprwc\" (UniqueName: \"kubernetes.io/projected/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-kube-api-access-cprwc\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.304031 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4925fb22-c2bb-4609-a593-5589e9137ecc-operator-scripts\") pod \"cinder-098a-account-create-7sqdp\" (UID: \"4925fb22-c2bb-4609-a593-5589e9137ecc\") " pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.323721 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp9rh\" (UniqueName: \"kubernetes.io/projected/4925fb22-c2bb-4609-a593-5589e9137ecc-kube-api-access-xp9rh\") pod \"cinder-098a-account-create-7sqdp\" (UID: \"4925fb22-c2bb-4609-a593-5589e9137ecc\") " pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.338053 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg-config-nl7bp" event={"ID":"d49383b2-a4dd-44d2-98e3-bc81f6d27a54","Type":"ContainerStarted","Data":"7da5b91310034032672fe15d0d6fcbd3a21951a2a004f33a8ad4d389ba9a4dd2"} Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.389475 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.404885 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-config-data\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.404948 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cprwc\" (UniqueName: \"kubernetes.io/projected/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-kube-api-access-cprwc\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.405069 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-combined-ca-bundle\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.413410 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-combined-ca-bundle\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.413549 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-config-data\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.431992 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.433167 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cprwc\" (UniqueName: \"kubernetes.io/projected/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-kube-api-access-cprwc\") pod \"keystone-db-sync-9grv6\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.448386 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.604650 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.604973 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="prometheus" containerID="cri-o://b3c9fc6ea8757266331eabc4a4429d89d80fb829b60d8538119de99da812b8ae" gracePeriod=600 Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.605124 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="thanos-sidecar" containerID="cri-o://37cc336eeb0e55dacf3617c3e737ce28e02bacb73aafccee97cb3bc4b4e0c6d2" gracePeriod=600 Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.605178 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="config-reloader" containerID="cri-o://33d4f40ef552d8b1c5dbed8c3b3d4e62283ddfc16a37f2822441144dcf24ecc5" gracePeriod=600 Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.666509 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-8brw8"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.667790 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.674591 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8brw8"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.707477 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-134f-account-create-g9cqb"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.709704 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.712633 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.715928 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9grv6" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.718512 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-134f-account-create-g9cqb"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.749329 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6rpd\" (UniqueName: \"kubernetes.io/projected/5a566e05-aacd-44f9-b3dd-231549b9a967-kube-api-access-x6rpd\") pod \"neutron-db-create-8brw8\" (UID: \"5a566e05-aacd-44f9-b3dd-231549b9a967\") " pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.749611 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhzt2\" (UniqueName: \"kubernetes.io/projected/c6331445-03fa-4581-a892-974fdd4e53fa-kube-api-access-bhzt2\") pod \"neutron-134f-account-create-g9cqb\" (UID: \"c6331445-03fa-4581-a892-974fdd4e53fa\") " pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.749695 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a566e05-aacd-44f9-b3dd-231549b9a967-operator-scripts\") pod \"neutron-db-create-8brw8\" (UID: \"5a566e05-aacd-44f9-b3dd-231549b9a967\") " pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.749843 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6331445-03fa-4581-a892-974fdd4e53fa-operator-scripts\") pod \"neutron-134f-account-create-g9cqb\" (UID: \"c6331445-03fa-4581-a892-974fdd4e53fa\") " pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.827518 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-t9t64"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.828889 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.829219 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-t9t64"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.849149 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-9cd9-account-create-zxr8b"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.851171 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwjt7\" (UniqueName: \"kubernetes.io/projected/422ffe22-0786-4688-bd88-9e28481c59ff-kube-api-access-zwjt7\") pod \"barbican-db-create-t9t64\" (UID: \"422ffe22-0786-4688-bd88-9e28481c59ff\") " pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.851218 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/422ffe22-0786-4688-bd88-9e28481c59ff-operator-scripts\") pod \"barbican-db-create-t9t64\" (UID: \"422ffe22-0786-4688-bd88-9e28481c59ff\") " pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.851256 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhzt2\" (UniqueName: \"kubernetes.io/projected/c6331445-03fa-4581-a892-974fdd4e53fa-kube-api-access-bhzt2\") pod \"neutron-134f-account-create-g9cqb\" (UID: \"c6331445-03fa-4581-a892-974fdd4e53fa\") " pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.851305 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a566e05-aacd-44f9-b3dd-231549b9a967-operator-scripts\") pod \"neutron-db-create-8brw8\" (UID: \"5a566e05-aacd-44f9-b3dd-231549b9a967\") " pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.851369 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6331445-03fa-4581-a892-974fdd4e53fa-operator-scripts\") pod \"neutron-134f-account-create-g9cqb\" (UID: \"c6331445-03fa-4581-a892-974fdd4e53fa\") " pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.851475 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6rpd\" (UniqueName: \"kubernetes.io/projected/5a566e05-aacd-44f9-b3dd-231549b9a967-kube-api-access-x6rpd\") pod \"neutron-db-create-8brw8\" (UID: \"5a566e05-aacd-44f9-b3dd-231549b9a967\") " pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.852520 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6331445-03fa-4581-a892-974fdd4e53fa-operator-scripts\") pod \"neutron-134f-account-create-g9cqb\" (UID: \"c6331445-03fa-4581-a892-974fdd4e53fa\") " pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.852559 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a566e05-aacd-44f9-b3dd-231549b9a967-operator-scripts\") pod \"neutron-db-create-8brw8\" (UID: \"5a566e05-aacd-44f9-b3dd-231549b9a967\") " pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.885196 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhzt2\" (UniqueName: \"kubernetes.io/projected/c6331445-03fa-4581-a892-974fdd4e53fa-kube-api-access-bhzt2\") pod \"neutron-134f-account-create-g9cqb\" (UID: \"c6331445-03fa-4581-a892-974fdd4e53fa\") " pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.894246 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6rpd\" (UniqueName: \"kubernetes.io/projected/5a566e05-aacd-44f9-b3dd-231549b9a967-kube-api-access-x6rpd\") pod \"neutron-db-create-8brw8\" (UID: \"5a566e05-aacd-44f9-b3dd-231549b9a967\") " pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:38 crc kubenswrapper[4735]: W1122 08:23:38.905993 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45fc366f_ea6e_4134_afa2_2612f5f0d5f3.slice/crio-f8dce2a6312beb903f5e58fe9a9736efc418183ceaef8f044b26fc676c58eb73 WatchSource:0}: Error finding container f8dce2a6312beb903f5e58fe9a9736efc418183ceaef8f044b26fc676c58eb73: Status 404 returned error can't find the container with id f8dce2a6312beb903f5e58fe9a9736efc418183ceaef8f044b26fc676c58eb73 Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.906322 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3454-account-create-j5ssm"] Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.907606 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.919173 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.952620 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-operator-scripts\") pod \"barbican-3454-account-create-j5ssm\" (UID: \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\") " pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.952672 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxbkq\" (UniqueName: \"kubernetes.io/projected/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-kube-api-access-nxbkq\") pod \"barbican-3454-account-create-j5ssm\" (UID: \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\") " pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.952805 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwjt7\" (UniqueName: \"kubernetes.io/projected/422ffe22-0786-4688-bd88-9e28481c59ff-kube-api-access-zwjt7\") pod \"barbican-db-create-t9t64\" (UID: \"422ffe22-0786-4688-bd88-9e28481c59ff\") " pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.952837 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/422ffe22-0786-4688-bd88-9e28481c59ff-operator-scripts\") pod \"barbican-db-create-t9t64\" (UID: \"422ffe22-0786-4688-bd88-9e28481c59ff\") " pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.953611 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/422ffe22-0786-4688-bd88-9e28481c59ff-operator-scripts\") pod \"barbican-db-create-t9t64\" (UID: \"422ffe22-0786-4688-bd88-9e28481c59ff\") " pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:38 crc kubenswrapper[4735]: I1122 08:23:38.989012 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwjt7\" (UniqueName: \"kubernetes.io/projected/422ffe22-0786-4688-bd88-9e28481c59ff-kube-api-access-zwjt7\") pod \"barbican-db-create-t9t64\" (UID: \"422ffe22-0786-4688-bd88-9e28481c59ff\") " pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.015987 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.038474 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.055124 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-operator-scripts\") pod \"barbican-3454-account-create-j5ssm\" (UID: \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\") " pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.055200 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxbkq\" (UniqueName: \"kubernetes.io/projected/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-kube-api-access-nxbkq\") pod \"barbican-3454-account-create-j5ssm\" (UID: \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\") " pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.056333 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-operator-scripts\") pod \"barbican-3454-account-create-j5ssm\" (UID: \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\") " pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.079549 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3454-account-create-j5ssm"] Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.089244 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxbkq\" (UniqueName: \"kubernetes.io/projected/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-kube-api-access-nxbkq\") pod \"barbican-3454-account-create-j5ssm\" (UID: \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\") " pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.190097 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.216194 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-bkgcz"] Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.272463 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.321778 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-7cghr"] Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.408671 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-098a-account-create-7sqdp"] Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.418323 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bkgcz" event={"ID":"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e","Type":"ContainerStarted","Data":"a6afe4c502addb6ad9ba899e980e2ac946fabf79b66b1cb35d83502afdf64b7b"} Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.420290 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-9cd9-account-create-zxr8b" event={"ID":"45fc366f-ea6e-4134-afa2-2612f5f0d5f3","Type":"ContainerStarted","Data":"f8dce2a6312beb903f5e58fe9a9736efc418183ceaef8f044b26fc676c58eb73"} Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.439673 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7cghr" event={"ID":"ebfceb98-899b-40d4-a27d-dfb27042557e","Type":"ContainerStarted","Data":"526046d4f8784d57ea838fae117ce152112cfaf781c3421d725b6ede641a358f"} Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.535734 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="prometheus" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 22 08:23:39 crc kubenswrapper[4735]: I1122 08:23:39.813589 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9grv6"] Nov 22 08:23:39 crc kubenswrapper[4735]: W1122 08:23:39.878245 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36d75e4f_e760_41d2_b7d8_b9baffb60bcf.slice/crio-9766c3b78da459609da1c207f9a2ccf206cb6007b7b78dd2e35bf5a500ca65db WatchSource:0}: Error finding container 9766c3b78da459609da1c207f9a2ccf206cb6007b7b78dd2e35bf5a500ca65db: Status 404 returned error can't find the container with id 9766c3b78da459609da1c207f9a2ccf206cb6007b7b78dd2e35bf5a500ca65db Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.043866 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8brw8"] Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.109943 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-t9t64"] Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.119406 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-134f-account-create-g9cqb"] Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.161765 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3454-account-create-j5ssm"] Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.450618 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9grv6" event={"ID":"36d75e4f-e760-41d2-b7d8-b9baffb60bcf","Type":"ContainerStarted","Data":"9766c3b78da459609da1c207f9a2ccf206cb6007b7b78dd2e35bf5a500ca65db"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.453335 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7cghr" event={"ID":"ebfceb98-899b-40d4-a27d-dfb27042557e","Type":"ContainerStarted","Data":"191e5ae3b96eca02d4554ac81a431b766af329bde7f9491675c3d30532a6b393"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.456063 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bkgcz" event={"ID":"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e","Type":"ContainerStarted","Data":"9d4cc87de46a3e403e4260565b03edabe0f0fe1931783dfc01b982dbd84a32bb"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.458625 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-9cd9-account-create-zxr8b" event={"ID":"45fc366f-ea6e-4134-afa2-2612f5f0d5f3","Type":"ContainerStarted","Data":"aeef6bab3bffb95e65aa4f8a85045f2ddbe112332a288077632042e0ab193708"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.461813 4735 generic.go:334] "Generic (PLEG): container finished" podID="bea30a30-b625-4511-87da-f26e5495a5f9" containerID="37cc336eeb0e55dacf3617c3e737ce28e02bacb73aafccee97cb3bc4b4e0c6d2" exitCode=0 Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.461843 4735 generic.go:334] "Generic (PLEG): container finished" podID="bea30a30-b625-4511-87da-f26e5495a5f9" containerID="33d4f40ef552d8b1c5dbed8c3b3d4e62283ddfc16a37f2822441144dcf24ecc5" exitCode=0 Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.461854 4735 generic.go:334] "Generic (PLEG): container finished" podID="bea30a30-b625-4511-87da-f26e5495a5f9" containerID="b3c9fc6ea8757266331eabc4a4429d89d80fb829b60d8538119de99da812b8ae" exitCode=0 Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.461922 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerDied","Data":"37cc336eeb0e55dacf3617c3e737ce28e02bacb73aafccee97cb3bc4b4e0c6d2"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.461967 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerDied","Data":"33d4f40ef552d8b1c5dbed8c3b3d4e62283ddfc16a37f2822441144dcf24ecc5"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.461982 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerDied","Data":"b3c9fc6ea8757266331eabc4a4429d89d80fb829b60d8538119de99da812b8ae"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.464765 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-098a-account-create-7sqdp" event={"ID":"4925fb22-c2bb-4609-a593-5589e9137ecc","Type":"ContainerStarted","Data":"df8ad461779002e80910d31d5dffe0fcbe435f4f1cde76d276ed7f11d6ae4d10"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.464813 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-098a-account-create-7sqdp" event={"ID":"4925fb22-c2bb-4609-a593-5589e9137ecc","Type":"ContainerStarted","Data":"2dbed7cf466da5ba5dc1db41d8666166aac187934b8453cae4f4cdec51085198"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.470084 4735 generic.go:334] "Generic (PLEG): container finished" podID="d49383b2-a4dd-44d2-98e3-bc81f6d27a54" containerID="83a6187f25b3c04cb4758ec8d289340f16e25e0b8c0847b4fd93fa50b6c4ca15" exitCode=0 Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.470142 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg-config-nl7bp" event={"ID":"d49383b2-a4dd-44d2-98e3-bc81f6d27a54","Type":"ContainerDied","Data":"83a6187f25b3c04cb4758ec8d289340f16e25e0b8c0847b4fd93fa50b6c4ca15"} Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.473690 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-7cghr" podStartSLOduration=3.473673051 podStartE2EDuration="3.473673051s" podCreationTimestamp="2025-11-22 08:23:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:40.469752252 +0000 UTC m=+1242.074090867" watchObservedRunningTime="2025-11-22 08:23:40.473673051 +0000 UTC m=+1242.078011656" Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.490400 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-9cd9-account-create-zxr8b" podStartSLOduration=3.4903829760000002 podStartE2EDuration="3.490382976s" podCreationTimestamp="2025-11-22 08:23:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:40.488827753 +0000 UTC m=+1242.093166378" watchObservedRunningTime="2025-11-22 08:23:40.490382976 +0000 UTC m=+1242.094721581" Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.506191 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-098a-account-create-7sqdp" podStartSLOduration=2.506158505 podStartE2EDuration="2.506158505s" podCreationTimestamp="2025-11-22 08:23:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:40.503993954 +0000 UTC m=+1242.108332559" watchObservedRunningTime="2025-11-22 08:23:40.506158505 +0000 UTC m=+1242.110497110" Nov 22 08:23:40 crc kubenswrapper[4735]: I1122 08:23:40.528988 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-bkgcz" podStartSLOduration=2.528957999 podStartE2EDuration="2.528957999s" podCreationTimestamp="2025-11-22 08:23:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:40.523823506 +0000 UTC m=+1242.128162111" watchObservedRunningTime="2025-11-22 08:23:40.528957999 +0000 UTC m=+1242.133296634" Nov 22 08:23:41 crc kubenswrapper[4735]: I1122 08:23:41.240856 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-z5csg" Nov 22 08:23:42 crc kubenswrapper[4735]: I1122 08:23:42.519113 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5ee793a0-c437-47a3-b05d-369a1730d6b1","Type":"ContainerStarted","Data":"c30125bdc4dbd76c5af54ee9db061d48b5996cf4d250fc32b82110f4c3dc513f"} Nov 22 08:23:43 crc kubenswrapper[4735]: I1122 08:23:43.527293 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:23:43 crc kubenswrapper[4735]: I1122 08:23:43.570505 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371959.284296 podStartE2EDuration="1m17.57048041s" podCreationTimestamp="2025-11-22 08:22:26 +0000 UTC" firstStartedPulling="2025-11-22 08:22:28.794033837 +0000 UTC m=+1170.398372442" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:43.555827943 +0000 UTC m=+1245.160166548" watchObservedRunningTime="2025-11-22 08:23:43.57048041 +0000 UTC m=+1245.174819025" Nov 22 08:23:44 crc kubenswrapper[4735]: I1122 08:23:44.538268 4735 generic.go:334] "Generic (PLEG): container finished" podID="ebfceb98-899b-40d4-a27d-dfb27042557e" containerID="191e5ae3b96eca02d4554ac81a431b766af329bde7f9491675c3d30532a6b393" exitCode=0 Nov 22 08:23:44 crc kubenswrapper[4735]: I1122 08:23:44.538351 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7cghr" event={"ID":"ebfceb98-899b-40d4-a27d-dfb27042557e","Type":"ContainerDied","Data":"191e5ae3b96eca02d4554ac81a431b766af329bde7f9491675c3d30532a6b393"} Nov 22 08:23:44 crc kubenswrapper[4735]: I1122 08:23:44.540544 4735 generic.go:334] "Generic (PLEG): container finished" podID="69db05e1-3ac1-4aaf-a558-6e8581dfbf2e" containerID="9d4cc87de46a3e403e4260565b03edabe0f0fe1931783dfc01b982dbd84a32bb" exitCode=0 Nov 22 08:23:44 crc kubenswrapper[4735]: I1122 08:23:44.540606 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bkgcz" event={"ID":"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e","Type":"ContainerDied","Data":"9d4cc87de46a3e403e4260565b03edabe0f0fe1931783dfc01b982dbd84a32bb"} Nov 22 08:23:44 crc kubenswrapper[4735]: I1122 08:23:44.542947 4735 generic.go:334] "Generic (PLEG): container finished" podID="45fc366f-ea6e-4134-afa2-2612f5f0d5f3" containerID="aeef6bab3bffb95e65aa4f8a85045f2ddbe112332a288077632042e0ab193708" exitCode=0 Nov 22 08:23:44 crc kubenswrapper[4735]: I1122 08:23:44.542993 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-9cd9-account-create-zxr8b" event={"ID":"45fc366f-ea6e-4134-afa2-2612f5f0d5f3","Type":"ContainerDied","Data":"aeef6bab3bffb95e65aa4f8a85045f2ddbe112332a288077632042e0ab193708"} Nov 22 08:23:44 crc kubenswrapper[4735]: I1122 08:23:44.545124 4735 generic.go:334] "Generic (PLEG): container finished" podID="4925fb22-c2bb-4609-a593-5589e9137ecc" containerID="df8ad461779002e80910d31d5dffe0fcbe435f4f1cde76d276ed7f11d6ae4d10" exitCode=0 Nov 22 08:23:44 crc kubenswrapper[4735]: I1122 08:23:44.545179 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-098a-account-create-7sqdp" event={"ID":"4925fb22-c2bb-4609-a593-5589e9137ecc","Type":"ContainerDied","Data":"df8ad461779002e80910d31d5dffe0fcbe435f4f1cde76d276ed7f11d6ae4d10"} Nov 22 08:23:46 crc kubenswrapper[4735]: I1122 08:23:46.132434 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:23:46 crc kubenswrapper[4735]: I1122 08:23:46.132846 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:23:47 crc kubenswrapper[4735]: I1122 08:23:47.534736 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.140:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:23:50 crc kubenswrapper[4735]: W1122 08:23:50.436322 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod422ffe22_0786_4688_bd88_9e28481c59ff.slice/crio-4c1c46007d9e821da4c4688dcccc56358f7a2b1ab7ab61b372a2c8db126cdd2a WatchSource:0}: Error finding container 4c1c46007d9e821da4c4688dcccc56358f7a2b1ab7ab61b372a2c8db126cdd2a: Status 404 returned error can't find the container with id 4c1c46007d9e821da4c4688dcccc56358f7a2b1ab7ab61b372a2c8db126cdd2a Nov 22 08:23:50 crc kubenswrapper[4735]: W1122 08:23:50.443705 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a566e05_aacd_44f9_b3dd_231549b9a967.slice/crio-4899c31f7cda37977ab6d743c8b57c165c8876ea7b2f363fb2f3e18c86d47b76 WatchSource:0}: Error finding container 4899c31f7cda37977ab6d743c8b57c165c8876ea7b2f363fb2f3e18c86d47b76: Status 404 returned error can't find the container with id 4899c31f7cda37977ab6d743c8b57c165c8876ea7b2f363fb2f3e18c86d47b76 Nov 22 08:23:50 crc kubenswrapper[4735]: W1122 08:23:50.449873 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6331445_03fa_4581_a892_974fdd4e53fa.slice/crio-a1dbd785d1332c6f70a1f35f967bc0d4479b19d1fa529d976186f47ec9ff7a09 WatchSource:0}: Error finding container a1dbd785d1332c6f70a1f35f967bc0d4479b19d1fa529d976186f47ec9ff7a09: Status 404 returned error can't find the container with id a1dbd785d1332c6f70a1f35f967bc0d4479b19d1fa529d976186f47ec9ff7a09 Nov 22 08:23:50 crc kubenswrapper[4735]: E1122 08:23:50.461796 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Nov 22 08:23:50 crc kubenswrapper[4735]: E1122 08:23:50.461985 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-twfqg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-m9x5r_openstack(54265b3d-7ff9-4b71-aab0-9b79ae0b08d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:23:50 crc kubenswrapper[4735]: E1122 08:23:50.463398 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-m9x5r" podUID="54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.619509 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-098a-account-create-7sqdp" event={"ID":"4925fb22-c2bb-4609-a593-5589e9137ecc","Type":"ContainerDied","Data":"2dbed7cf466da5ba5dc1db41d8666166aac187934b8453cae4f4cdec51085198"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.619559 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dbed7cf466da5ba5dc1db41d8666166aac187934b8453cae4f4cdec51085198" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.622901 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg-config-nl7bp" event={"ID":"d49383b2-a4dd-44d2-98e3-bc81f6d27a54","Type":"ContainerDied","Data":"7da5b91310034032672fe15d0d6fcbd3a21951a2a004f33a8ad4d389ba9a4dd2"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.622945 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7da5b91310034032672fe15d0d6fcbd3a21951a2a004f33a8ad4d389ba9a4dd2" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.626414 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bkgcz" event={"ID":"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e","Type":"ContainerDied","Data":"a6afe4c502addb6ad9ba899e980e2ac946fabf79b66b1cb35d83502afdf64b7b"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.627015 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6afe4c502addb6ad9ba899e980e2ac946fabf79b66b1cb35d83502afdf64b7b" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.628350 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8brw8" event={"ID":"5a566e05-aacd-44f9-b3dd-231549b9a967","Type":"ContainerStarted","Data":"4899c31f7cda37977ab6d743c8b57c165c8876ea7b2f363fb2f3e18c86d47b76"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.643769 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"bea30a30-b625-4511-87da-f26e5495a5f9","Type":"ContainerDied","Data":"85f8d0995934a219cb503ddae2705f2d58f00c911193c039df90ff7cc690d342"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.643821 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85f8d0995934a219cb503ddae2705f2d58f00c911193c039df90ff7cc690d342" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.646136 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3454-account-create-j5ssm" event={"ID":"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7","Type":"ContainerStarted","Data":"74bde05ac6b17126eb06e854cc094b4b26150fb38c02537a7950ae833eced389"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.648185 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7cghr" event={"ID":"ebfceb98-899b-40d4-a27d-dfb27042557e","Type":"ContainerDied","Data":"526046d4f8784d57ea838fae117ce152112cfaf781c3421d725b6ede641a358f"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.648215 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="526046d4f8784d57ea838fae117ce152112cfaf781c3421d725b6ede641a358f" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.649570 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-9cd9-account-create-zxr8b" event={"ID":"45fc366f-ea6e-4134-afa2-2612f5f0d5f3","Type":"ContainerDied","Data":"f8dce2a6312beb903f5e58fe9a9736efc418183ceaef8f044b26fc676c58eb73"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.649595 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8dce2a6312beb903f5e58fe9a9736efc418183ceaef8f044b26fc676c58eb73" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.651998 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-134f-account-create-g9cqb" event={"ID":"c6331445-03fa-4581-a892-974fdd4e53fa","Type":"ContainerStarted","Data":"a1dbd785d1332c6f70a1f35f967bc0d4479b19d1fa529d976186f47ec9ff7a09"} Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.654101 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t9t64" event={"ID":"422ffe22-0786-4688-bd88-9e28481c59ff","Type":"ContainerStarted","Data":"4c1c46007d9e821da4c4688dcccc56358f7a2b1ab7ab61b372a2c8db126cdd2a"} Nov 22 08:23:50 crc kubenswrapper[4735]: E1122 08:23:50.656877 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-m9x5r" podUID="54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.668400 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.763895 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.793999 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.795572 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf6x2\" (UniqueName: \"kubernetes.io/projected/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-kube-api-access-bf6x2\") pod \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\" (UID: \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\") " Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.796572 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-operator-scripts\") pod \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\" (UID: \"69db05e1-3ac1-4aaf-a558-6e8581dfbf2e\") " Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.798252 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "69db05e1-3ac1-4aaf-a558-6e8581dfbf2e" (UID: "69db05e1-3ac1-4aaf-a558-6e8581dfbf2e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.801548 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.822967 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-kube-api-access-bf6x2" (OuterVolumeSpecName: "kube-api-access-bf6x2") pod "69db05e1-3ac1-4aaf-a558-6e8581dfbf2e" (UID: "69db05e1-3ac1-4aaf-a558-6e8581dfbf2e"). InnerVolumeSpecName "kube-api-access-bf6x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.866315 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.898438 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-operator-scripts\") pod \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\" (UID: \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\") " Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.898520 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebfceb98-899b-40d4-a27d-dfb27042557e-operator-scripts\") pod \"ebfceb98-899b-40d4-a27d-dfb27042557e\" (UID: \"ebfceb98-899b-40d4-a27d-dfb27042557e\") " Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.898560 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4925fb22-c2bb-4609-a593-5589e9137ecc-operator-scripts\") pod \"4925fb22-c2bb-4609-a593-5589e9137ecc\" (UID: \"4925fb22-c2bb-4609-a593-5589e9137ecc\") " Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.898725 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9zxr\" (UniqueName: \"kubernetes.io/projected/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-kube-api-access-x9zxr\") pod \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\" (UID: \"45fc366f-ea6e-4134-afa2-2612f5f0d5f3\") " Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.898821 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp9rh\" (UniqueName: \"kubernetes.io/projected/4925fb22-c2bb-4609-a593-5589e9137ecc-kube-api-access-xp9rh\") pod \"4925fb22-c2bb-4609-a593-5589e9137ecc\" (UID: \"4925fb22-c2bb-4609-a593-5589e9137ecc\") " Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.898877 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjxjx\" (UniqueName: \"kubernetes.io/projected/ebfceb98-899b-40d4-a27d-dfb27042557e-kube-api-access-xjxjx\") pod \"ebfceb98-899b-40d4-a27d-dfb27042557e\" (UID: \"ebfceb98-899b-40d4-a27d-dfb27042557e\") " Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.899359 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.899378 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf6x2\" (UniqueName: \"kubernetes.io/projected/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e-kube-api-access-bf6x2\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.901831 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4925fb22-c2bb-4609-a593-5589e9137ecc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4925fb22-c2bb-4609-a593-5589e9137ecc" (UID: "4925fb22-c2bb-4609-a593-5589e9137ecc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.902859 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "45fc366f-ea6e-4134-afa2-2612f5f0d5f3" (UID: "45fc366f-ea6e-4134-afa2-2612f5f0d5f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.904958 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebfceb98-899b-40d4-a27d-dfb27042557e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ebfceb98-899b-40d4-a27d-dfb27042557e" (UID: "ebfceb98-899b-40d4-a27d-dfb27042557e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.907366 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebfceb98-899b-40d4-a27d-dfb27042557e-kube-api-access-xjxjx" (OuterVolumeSpecName: "kube-api-access-xjxjx") pod "ebfceb98-899b-40d4-a27d-dfb27042557e" (UID: "ebfceb98-899b-40d4-a27d-dfb27042557e"). InnerVolumeSpecName "kube-api-access-xjxjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.914755 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-kube-api-access-x9zxr" (OuterVolumeSpecName: "kube-api-access-x9zxr") pod "45fc366f-ea6e-4134-afa2-2612f5f0d5f3" (UID: "45fc366f-ea6e-4134-afa2-2612f5f0d5f3"). InnerVolumeSpecName "kube-api-access-x9zxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.914826 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4925fb22-c2bb-4609-a593-5589e9137ecc-kube-api-access-xp9rh" (OuterVolumeSpecName: "kube-api-access-xp9rh") pod "4925fb22-c2bb-4609-a593-5589e9137ecc" (UID: "4925fb22-c2bb-4609-a593-5589e9137ecc"). InnerVolumeSpecName "kube-api-access-xp9rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:50 crc kubenswrapper[4735]: I1122 08:23:50.919308 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.000772 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/bea30a30-b625-4511-87da-f26e5495a5f9-prometheus-metric-storage-rulefiles-0\") pod \"bea30a30-b625-4511-87da-f26e5495a5f9\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.000855 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-config\") pod \"bea30a30-b625-4511-87da-f26e5495a5f9\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.001591 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9jkq\" (UniqueName: \"kubernetes.io/projected/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-kube-api-access-l9jkq\") pod \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.001640 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run-ovn\") pod \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.001665 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-web-config\") pod \"bea30a30-b625-4511-87da-f26e5495a5f9\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.001787 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"bea30a30-b625-4511-87da-f26e5495a5f9\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.001862 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run\") pod \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.001892 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-thanos-prometheus-http-client-file\") pod \"bea30a30-b625-4511-87da-f26e5495a5f9\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.001951 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-log-ovn\") pod \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.002017 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-additional-scripts\") pod \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.002043 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-scripts\") pod \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\" (UID: \"d49383b2-a4dd-44d2-98e3-bc81f6d27a54\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.002047 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bea30a30-b625-4511-87da-f26e5495a5f9-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "bea30a30-b625-4511-87da-f26e5495a5f9" (UID: "bea30a30-b625-4511-87da-f26e5495a5f9"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.002096 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/bea30a30-b625-4511-87da-f26e5495a5f9-config-out\") pod \"bea30a30-b625-4511-87da-f26e5495a5f9\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.002114 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run" (OuterVolumeSpecName: "var-run") pod "d49383b2-a4dd-44d2-98e3-bc81f6d27a54" (UID: "d49383b2-a4dd-44d2-98e3-bc81f6d27a54"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.002116 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-tls-assets\") pod \"bea30a30-b625-4511-87da-f26e5495a5f9\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.002184 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj4ll\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-kube-api-access-jj4ll\") pod \"bea30a30-b625-4511-87da-f26e5495a5f9\" (UID: \"bea30a30-b625-4511-87da-f26e5495a5f9\") " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003119 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9zxr\" (UniqueName: \"kubernetes.io/projected/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-kube-api-access-x9zxr\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003135 4735 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/bea30a30-b625-4511-87da-f26e5495a5f9-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003149 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp9rh\" (UniqueName: \"kubernetes.io/projected/4925fb22-c2bb-4609-a593-5589e9137ecc-kube-api-access-xp9rh\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003158 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjxjx\" (UniqueName: \"kubernetes.io/projected/ebfceb98-899b-40d4-a27d-dfb27042557e-kube-api-access-xjxjx\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003167 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45fc366f-ea6e-4134-afa2-2612f5f0d5f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003176 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebfceb98-899b-40d4-a27d-dfb27042557e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003187 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4925fb22-c2bb-4609-a593-5589e9137ecc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003196 4735 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003801 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d49383b2-a4dd-44d2-98e3-bc81f6d27a54" (UID: "d49383b2-a4dd-44d2-98e3-bc81f6d27a54"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.003865 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d49383b2-a4dd-44d2-98e3-bc81f6d27a54" (UID: "d49383b2-a4dd-44d2-98e3-bc81f6d27a54"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.006335 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d49383b2-a4dd-44d2-98e3-bc81f6d27a54" (UID: "d49383b2-a4dd-44d2-98e3-bc81f6d27a54"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.006445 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-kube-api-access-l9jkq" (OuterVolumeSpecName: "kube-api-access-l9jkq") pod "d49383b2-a4dd-44d2-98e3-bc81f6d27a54" (UID: "d49383b2-a4dd-44d2-98e3-bc81f6d27a54"). InnerVolumeSpecName "kube-api-access-l9jkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.007914 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-config" (OuterVolumeSpecName: "config") pod "bea30a30-b625-4511-87da-f26e5495a5f9" (UID: "bea30a30-b625-4511-87da-f26e5495a5f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.007986 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "bea30a30-b625-4511-87da-f26e5495a5f9" (UID: "bea30a30-b625-4511-87da-f26e5495a5f9"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.010114 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "bea30a30-b625-4511-87da-f26e5495a5f9" (UID: "bea30a30-b625-4511-87da-f26e5495a5f9"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.011938 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-scripts" (OuterVolumeSpecName: "scripts") pod "d49383b2-a4dd-44d2-98e3-bc81f6d27a54" (UID: "d49383b2-a4dd-44d2-98e3-bc81f6d27a54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.016029 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bea30a30-b625-4511-87da-f26e5495a5f9-config-out" (OuterVolumeSpecName: "config-out") pod "bea30a30-b625-4511-87da-f26e5495a5f9" (UID: "bea30a30-b625-4511-87da-f26e5495a5f9"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.018956 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-kube-api-access-jj4ll" (OuterVolumeSpecName: "kube-api-access-jj4ll") pod "bea30a30-b625-4511-87da-f26e5495a5f9" (UID: "bea30a30-b625-4511-87da-f26e5495a5f9"). InnerVolumeSpecName "kube-api-access-jj4ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.032452 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "bea30a30-b625-4511-87da-f26e5495a5f9" (UID: "bea30a30-b625-4511-87da-f26e5495a5f9"). InnerVolumeSpecName "pvc-a7945675-c855-462a-adda-9313711660b4". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.048180 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-web-config" (OuterVolumeSpecName: "web-config") pod "bea30a30-b625-4511-87da-f26e5495a5f9" (UID: "bea30a30-b625-4511-87da-f26e5495a5f9"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110848 4735 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110893 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110907 4735 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/bea30a30-b625-4511-87da-f26e5495a5f9-config-out\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110920 4735 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110932 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj4ll\" (UniqueName: \"kubernetes.io/projected/bea30a30-b625-4511-87da-f26e5495a5f9-kube-api-access-jj4ll\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110946 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110958 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9jkq\" (UniqueName: \"kubernetes.io/projected/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-kube-api-access-l9jkq\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110971 4735 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.110983 4735 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-web-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.111024 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") on node \"crc\" " Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.111040 4735 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/bea30a30-b625-4511-87da-f26e5495a5f9-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.111053 4735 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d49383b2-a4dd-44d2-98e3-bc81f6d27a54-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.181142 4735 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.181312 4735 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-a7945675-c855-462a-adda-9313711660b4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4") on node "crc" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.213053 4735 reconciler_common.go:293] "Volume detached for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.666830 4735 generic.go:334] "Generic (PLEG): container finished" podID="c6331445-03fa-4581-a892-974fdd4e53fa" containerID="69c97d42e592285727803bccf2ff3a143a24f89af4233304a34858d0578bf42a" exitCode=0 Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.666923 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-134f-account-create-g9cqb" event={"ID":"c6331445-03fa-4581-a892-974fdd4e53fa","Type":"ContainerDied","Data":"69c97d42e592285727803bccf2ff3a143a24f89af4233304a34858d0578bf42a"} Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.669100 4735 generic.go:334] "Generic (PLEG): container finished" podID="5a566e05-aacd-44f9-b3dd-231549b9a967" containerID="4856d1509ff777af929d217195a1f000a5c6a0ea2ac36c72a57d4e47b30fd734" exitCode=0 Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.669197 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8brw8" event={"ID":"5a566e05-aacd-44f9-b3dd-231549b9a967","Type":"ContainerDied","Data":"4856d1509ff777af929d217195a1f000a5c6a0ea2ac36c72a57d4e47b30fd734"} Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.674933 4735 generic.go:334] "Generic (PLEG): container finished" podID="422ffe22-0786-4688-bd88-9e28481c59ff" containerID="3f3c694cb46e4f85d2d5c053df886a6d1f1e4b54dd01578dbecbe67748a0d446" exitCode=0 Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.674983 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t9t64" event={"ID":"422ffe22-0786-4688-bd88-9e28481c59ff","Type":"ContainerDied","Data":"3f3c694cb46e4f85d2d5c053df886a6d1f1e4b54dd01578dbecbe67748a0d446"} Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.677491 4735 generic.go:334] "Generic (PLEG): container finished" podID="8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7" containerID="6ed26bbccf97073cf14b2cfbbe6e9da1892c4f89b3bc9b736a52aaa8d5d986f5" exitCode=0 Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.677600 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.677630 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3454-account-create-j5ssm" event={"ID":"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7","Type":"ContainerDied","Data":"6ed26bbccf97073cf14b2cfbbe6e9da1892c4f89b3bc9b736a52aaa8d5d986f5"} Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.678098 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-098a-account-create-7sqdp" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.678169 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-9cd9-account-create-zxr8b" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.678179 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bkgcz" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.678211 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7cghr" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.678189 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg-config-nl7bp" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.811219 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.837949 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.860346 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.860990 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="prometheus" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861011 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="prometheus" Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.861039 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="config-reloader" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861050 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="config-reloader" Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.861064 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69db05e1-3ac1-4aaf-a558-6e8581dfbf2e" containerName="mariadb-database-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861072 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="69db05e1-3ac1-4aaf-a558-6e8581dfbf2e" containerName="mariadb-database-create" Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.861098 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebfceb98-899b-40d4-a27d-dfb27042557e" containerName="mariadb-database-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861105 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebfceb98-899b-40d4-a27d-dfb27042557e" containerName="mariadb-database-create" Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.861123 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45fc366f-ea6e-4134-afa2-2612f5f0d5f3" containerName="mariadb-account-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861131 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="45fc366f-ea6e-4134-afa2-2612f5f0d5f3" containerName="mariadb-account-create" Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.861147 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="init-config-reloader" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861155 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="init-config-reloader" Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.861167 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="thanos-sidecar" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861174 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="thanos-sidecar" Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.861207 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49383b2-a4dd-44d2-98e3-bc81f6d27a54" containerName="ovn-config" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861215 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49383b2-a4dd-44d2-98e3-bc81f6d27a54" containerName="ovn-config" Nov 22 08:23:51 crc kubenswrapper[4735]: E1122 08:23:51.861231 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4925fb22-c2bb-4609-a593-5589e9137ecc" containerName="mariadb-account-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861239 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4925fb22-c2bb-4609-a593-5589e9137ecc" containerName="mariadb-account-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861502 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4925fb22-c2bb-4609-a593-5589e9137ecc" containerName="mariadb-account-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861520 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="config-reloader" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861535 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="thanos-sidecar" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861549 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49383b2-a4dd-44d2-98e3-bc81f6d27a54" containerName="ovn-config" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861571 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="45fc366f-ea6e-4134-afa2-2612f5f0d5f3" containerName="mariadb-account-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861585 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="prometheus" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861599 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="69db05e1-3ac1-4aaf-a558-6e8581dfbf2e" containerName="mariadb-database-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.861610 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebfceb98-899b-40d4-a27d-dfb27042557e" containerName="mariadb-database-create" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.863982 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.869594 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-gd8n8" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.869668 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.869853 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.869889 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.870365 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.871218 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.881568 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.883500 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927170 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqmgw\" (UniqueName: \"kubernetes.io/projected/f1d9efee-4efc-460c-ba99-36df95d96a48-kube-api-access-kqmgw\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927240 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f1d9efee-4efc-460c-ba99-36df95d96a48-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927287 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927343 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f1d9efee-4efc-460c-ba99-36df95d96a48-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927376 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927420 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927446 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927510 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-config\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927541 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927597 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f1d9efee-4efc-460c-ba99-36df95d96a48-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.927622 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.984163 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-z5csg-config-nl7bp"] Nov 22 08:23:51 crc kubenswrapper[4735]: I1122 08:23:51.995086 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-z5csg-config-nl7bp"] Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.028947 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f1d9efee-4efc-460c-ba99-36df95d96a48-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029036 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029109 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f1d9efee-4efc-460c-ba99-36df95d96a48-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029154 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029208 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029234 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029282 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-config\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029310 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029406 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f1d9efee-4efc-460c-ba99-36df95d96a48-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029437 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.029565 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqmgw\" (UniqueName: \"kubernetes.io/projected/f1d9efee-4efc-460c-ba99-36df95d96a48-kube-api-access-kqmgw\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.035487 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f1d9efee-4efc-460c-ba99-36df95d96a48-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.036239 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.036712 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f1d9efee-4efc-460c-ba99-36df95d96a48-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.037005 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-config\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.038210 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f1d9efee-4efc-460c-ba99-36df95d96a48-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.042169 4735 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.042206 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/504ad61e9b31def24aaaccbb19a5701f8cf89a1c3574104486b5c4b008bcacd8/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.042540 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.042698 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.043073 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.046606 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f1d9efee-4efc-460c-ba99-36df95d96a48-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.066204 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z5csg-config-m6z5s"] Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.067670 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.076931 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.086424 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqmgw\" (UniqueName: \"kubernetes.io/projected/f1d9efee-4efc-460c-ba99-36df95d96a48-kube-api-access-kqmgw\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.098309 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z5csg-config-m6z5s"] Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.119865 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a7945675-c855-462a-adda-9313711660b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7945675-c855-462a-adda-9313711660b4\") pod \"prometheus-metric-storage-0\" (UID: \"f1d9efee-4efc-460c-ba99-36df95d96a48\") " pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.130884 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-scripts\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.131088 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-additional-scripts\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.131139 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run-ovn\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.131185 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t4l7\" (UniqueName: \"kubernetes.io/projected/e41e549f-4050-4d3e-bad3-71e3ade56f86-kube-api-access-9t4l7\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.131216 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.131243 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-log-ovn\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.202676 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257134 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-additional-scripts\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257201 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run-ovn\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257242 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t4l7\" (UniqueName: \"kubernetes.io/projected/e41e549f-4050-4d3e-bad3-71e3ade56f86-kube-api-access-9t4l7\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257262 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257286 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-log-ovn\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257372 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-scripts\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257617 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run-ovn\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257610 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.257717 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-log-ovn\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.258027 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-additional-scripts\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.259394 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-scripts\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.289403 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t4l7\" (UniqueName: \"kubernetes.io/projected/e41e549f-4050-4d3e-bad3-71e3ade56f86-kube-api-access-9t4l7\") pod \"ovn-controller-z5csg-config-m6z5s\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.468001 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:23:52 crc kubenswrapper[4735]: I1122 08:23:52.534891 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.140:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.145687 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.274935 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bea30a30-b625-4511-87da-f26e5495a5f9" path="/var/lib/kubelet/pods/bea30a30-b625-4511-87da-f26e5495a5f9/volumes" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.275654 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49383b2-a4dd-44d2-98e3-bc81f6d27a54" path="/var/lib/kubelet/pods/d49383b2-a4dd-44d2-98e3-bc81f6d27a54/volumes" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.277039 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhzt2\" (UniqueName: \"kubernetes.io/projected/c6331445-03fa-4581-a892-974fdd4e53fa-kube-api-access-bhzt2\") pod \"c6331445-03fa-4581-a892-974fdd4e53fa\" (UID: \"c6331445-03fa-4581-a892-974fdd4e53fa\") " Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.277133 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6331445-03fa-4581-a892-974fdd4e53fa-operator-scripts\") pod \"c6331445-03fa-4581-a892-974fdd4e53fa\" (UID: \"c6331445-03fa-4581-a892-974fdd4e53fa\") " Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.278510 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6331445-03fa-4581-a892-974fdd4e53fa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c6331445-03fa-4581-a892-974fdd4e53fa" (UID: "c6331445-03fa-4581-a892-974fdd4e53fa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.288935 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6331445-03fa-4581-a892-974fdd4e53fa-kube-api-access-bhzt2" (OuterVolumeSpecName: "kube-api-access-bhzt2") pod "c6331445-03fa-4581-a892-974fdd4e53fa" (UID: "c6331445-03fa-4581-a892-974fdd4e53fa"). InnerVolumeSpecName "kube-api-access-bhzt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.380396 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhzt2\" (UniqueName: \"kubernetes.io/projected/c6331445-03fa-4581-a892-974fdd4e53fa-kube-api-access-bhzt2\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.380428 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6331445-03fa-4581-a892-974fdd4e53fa-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.500334 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z5csg-config-m6z5s"] Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.505476 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.506977 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.525868 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.584516 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.586740 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6rpd\" (UniqueName: \"kubernetes.io/projected/5a566e05-aacd-44f9-b3dd-231549b9a967-kube-api-access-x6rpd\") pod \"5a566e05-aacd-44f9-b3dd-231549b9a967\" (UID: \"5a566e05-aacd-44f9-b3dd-231549b9a967\") " Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.586821 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxbkq\" (UniqueName: \"kubernetes.io/projected/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-kube-api-access-nxbkq\") pod \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\" (UID: \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\") " Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.586920 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a566e05-aacd-44f9-b3dd-231549b9a967-operator-scripts\") pod \"5a566e05-aacd-44f9-b3dd-231549b9a967\" (UID: \"5a566e05-aacd-44f9-b3dd-231549b9a967\") " Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.586955 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/422ffe22-0786-4688-bd88-9e28481c59ff-operator-scripts\") pod \"422ffe22-0786-4688-bd88-9e28481c59ff\" (UID: \"422ffe22-0786-4688-bd88-9e28481c59ff\") " Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.587074 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwjt7\" (UniqueName: \"kubernetes.io/projected/422ffe22-0786-4688-bd88-9e28481c59ff-kube-api-access-zwjt7\") pod \"422ffe22-0786-4688-bd88-9e28481c59ff\" (UID: \"422ffe22-0786-4688-bd88-9e28481c59ff\") " Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.587132 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-operator-scripts\") pod \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\" (UID: \"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7\") " Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.588339 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7" (UID: "8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.592035 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a566e05-aacd-44f9-b3dd-231549b9a967-kube-api-access-x6rpd" (OuterVolumeSpecName: "kube-api-access-x6rpd") pod "5a566e05-aacd-44f9-b3dd-231549b9a967" (UID: "5a566e05-aacd-44f9-b3dd-231549b9a967"). InnerVolumeSpecName "kube-api-access-x6rpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.592377 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/422ffe22-0786-4688-bd88-9e28481c59ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "422ffe22-0786-4688-bd88-9e28481c59ff" (UID: "422ffe22-0786-4688-bd88-9e28481c59ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.592745 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a566e05-aacd-44f9-b3dd-231549b9a967-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a566e05-aacd-44f9-b3dd-231549b9a967" (UID: "5a566e05-aacd-44f9-b3dd-231549b9a967"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:23:53 crc kubenswrapper[4735]: W1122 08:23:53.594903 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1d9efee_4efc_460c_ba99_36df95d96a48.slice/crio-dd332cb0f7f518cfc3b0a2f54a6e6231727e33cb5d4071dcf7878180dc59ffb0 WatchSource:0}: Error finding container dd332cb0f7f518cfc3b0a2f54a6e6231727e33cb5d4071dcf7878180dc59ffb0: Status 404 returned error can't find the container with id dd332cb0f7f518cfc3b0a2f54a6e6231727e33cb5d4071dcf7878180dc59ffb0 Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.598033 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-kube-api-access-nxbkq" (OuterVolumeSpecName: "kube-api-access-nxbkq") pod "8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7" (UID: "8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7"). InnerVolumeSpecName "kube-api-access-nxbkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.604420 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/422ffe22-0786-4688-bd88-9e28481c59ff-kube-api-access-zwjt7" (OuterVolumeSpecName: "kube-api-access-zwjt7") pod "422ffe22-0786-4688-bd88-9e28481c59ff" (UID: "422ffe22-0786-4688-bd88-9e28481c59ff"). InnerVolumeSpecName "kube-api-access-zwjt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.689348 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6rpd\" (UniqueName: \"kubernetes.io/projected/5a566e05-aacd-44f9-b3dd-231549b9a967-kube-api-access-x6rpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.689377 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxbkq\" (UniqueName: \"kubernetes.io/projected/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-kube-api-access-nxbkq\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.689387 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a566e05-aacd-44f9-b3dd-231549b9a967-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.689397 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/422ffe22-0786-4688-bd88-9e28481c59ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.689435 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwjt7\" (UniqueName: \"kubernetes.io/projected/422ffe22-0786-4688-bd88-9e28481c59ff-kube-api-access-zwjt7\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.689444 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.704036 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"3b29201e-860c-4944-ad29-06abb7aed0c6","Type":"ContainerStarted","Data":"12ecbfb69646afa24be66e782ce2df4e41d0007a79333f6206eaf9333ca628a4"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.707202 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t9t64" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.707203 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t9t64" event={"ID":"422ffe22-0786-4688-bd88-9e28481c59ff","Type":"ContainerDied","Data":"4c1c46007d9e821da4c4688dcccc56358f7a2b1ab7ab61b372a2c8db126cdd2a"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.707245 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c1c46007d9e821da4c4688dcccc56358f7a2b1ab7ab61b372a2c8db126cdd2a" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.711785 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"44e85be6efb7ee968ca6d2b555af348c33337660898a04a8fab93f0fd6719523"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.711826 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"69f1bba3a771e5c2f0a90064042ee9030bd0f4751a9b7856a22da815c8bbc6ca"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.711836 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"54fed266eee9e5114f088abbe037cb6490024eaff42f601224f59cecc77c1909"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.714361 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3454-account-create-j5ssm" event={"ID":"8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7","Type":"ContainerDied","Data":"74bde05ac6b17126eb06e854cc094b4b26150fb38c02537a7950ae833eced389"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.714392 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74bde05ac6b17126eb06e854cc094b4b26150fb38c02537a7950ae833eced389" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.714415 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3454-account-create-j5ssm" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.716111 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f1d9efee-4efc-460c-ba99-36df95d96a48","Type":"ContainerStarted","Data":"dd332cb0f7f518cfc3b0a2f54a6e6231727e33cb5d4071dcf7878180dc59ffb0"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.717388 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg-config-m6z5s" event={"ID":"e41e549f-4050-4d3e-bad3-71e3ade56f86","Type":"ContainerStarted","Data":"fa58a9366e0a998c0b244b633ba8fe741812c12df39c77ddb6b202bf854eda81"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.728652 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-134f-account-create-g9cqb" event={"ID":"c6331445-03fa-4581-a892-974fdd4e53fa","Type":"ContainerDied","Data":"a1dbd785d1332c6f70a1f35f967bc0d4479b19d1fa529d976186f47ec9ff7a09"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.728690 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1dbd785d1332c6f70a1f35f967bc0d4479b19d1fa529d976186f47ec9ff7a09" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.728777 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-134f-account-create-g9cqb" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.744110 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8brw8" event={"ID":"5a566e05-aacd-44f9-b3dd-231549b9a967","Type":"ContainerDied","Data":"4899c31f7cda37977ab6d743c8b57c165c8876ea7b2f363fb2f3e18c86d47b76"} Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.744145 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4899c31f7cda37977ab6d743c8b57c165c8876ea7b2f363fb2f3e18c86d47b76" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.744199 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8brw8" Nov 22 08:23:53 crc kubenswrapper[4735]: I1122 08:23:53.758346 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.608445899 podStartE2EDuration="20.758327086s" podCreationTimestamp="2025-11-22 08:23:33 +0000 UTC" firstStartedPulling="2025-11-22 08:23:34.604666501 +0000 UTC m=+1236.209005116" lastFinishedPulling="2025-11-22 08:23:52.754547698 +0000 UTC m=+1254.358886303" observedRunningTime="2025-11-22 08:23:53.726401007 +0000 UTC m=+1255.330739612" watchObservedRunningTime="2025-11-22 08:23:53.758327086 +0000 UTC m=+1255.362665691" Nov 22 08:23:54 crc kubenswrapper[4735]: I1122 08:23:54.758441 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg-config-m6z5s" event={"ID":"e41e549f-4050-4d3e-bad3-71e3ade56f86","Type":"ContainerStarted","Data":"ca5814fdbfd4c6c36249028f417d2a8534621e0cd7a9496af74de53682f2201c"} Nov 22 08:23:54 crc kubenswrapper[4735]: I1122 08:23:54.762323 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"1226dc6e38a8f4275a4a3222280f90110c1c6e7d9cb728027212bc936f4741bb"} Nov 22 08:23:54 crc kubenswrapper[4735]: I1122 08:23:54.783850 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-z5csg-config-m6z5s" podStartSLOduration=2.783827707 podStartE2EDuration="2.783827707s" podCreationTimestamp="2025-11-22 08:23:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:23:54.774326932 +0000 UTC m=+1256.378665557" watchObservedRunningTime="2025-11-22 08:23:54.783827707 +0000 UTC m=+1256.388166312" Nov 22 08:23:55 crc kubenswrapper[4735]: I1122 08:23:55.776972 4735 generic.go:334] "Generic (PLEG): container finished" podID="e41e549f-4050-4d3e-bad3-71e3ade56f86" containerID="ca5814fdbfd4c6c36249028f417d2a8534621e0cd7a9496af74de53682f2201c" exitCode=0 Nov 22 08:23:55 crc kubenswrapper[4735]: I1122 08:23:55.777294 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg-config-m6z5s" event={"ID":"e41e549f-4050-4d3e-bad3-71e3ade56f86","Type":"ContainerDied","Data":"ca5814fdbfd4c6c36249028f417d2a8534621e0cd7a9496af74de53682f2201c"} Nov 22 08:23:57 crc kubenswrapper[4735]: I1122 08:23:57.805379 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f1d9efee-4efc-460c-ba99-36df95d96a48","Type":"ContainerStarted","Data":"264d532f6ae9cd12781153d0dcb50e74bd0c3beba046296f86e4bbb734e78963"} Nov 22 08:23:58 crc kubenswrapper[4735]: I1122 08:23:58.042778 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.278408 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.452410 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run-ovn\") pod \"e41e549f-4050-4d3e-bad3-71e3ade56f86\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.452782 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-log-ovn\") pod \"e41e549f-4050-4d3e-bad3-71e3ade56f86\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.452780 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e41e549f-4050-4d3e-bad3-71e3ade56f86" (UID: "e41e549f-4050-4d3e-bad3-71e3ade56f86"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.452844 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run\") pod \"e41e549f-4050-4d3e-bad3-71e3ade56f86\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.452884 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e41e549f-4050-4d3e-bad3-71e3ade56f86" (UID: "e41e549f-4050-4d3e-bad3-71e3ade56f86"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.452919 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run" (OuterVolumeSpecName: "var-run") pod "e41e549f-4050-4d3e-bad3-71e3ade56f86" (UID: "e41e549f-4050-4d3e-bad3-71e3ade56f86"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.452898 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-scripts\") pod \"e41e549f-4050-4d3e-bad3-71e3ade56f86\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.453335 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t4l7\" (UniqueName: \"kubernetes.io/projected/e41e549f-4050-4d3e-bad3-71e3ade56f86-kube-api-access-9t4l7\") pod \"e41e549f-4050-4d3e-bad3-71e3ade56f86\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.453365 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-additional-scripts\") pod \"e41e549f-4050-4d3e-bad3-71e3ade56f86\" (UID: \"e41e549f-4050-4d3e-bad3-71e3ade56f86\") " Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.454348 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e41e549f-4050-4d3e-bad3-71e3ade56f86" (UID: "e41e549f-4050-4d3e-bad3-71e3ade56f86"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.455045 4735 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.455065 4735 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.455074 4735 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e41e549f-4050-4d3e-bad3-71e3ade56f86-var-run\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.455081 4735 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.455192 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-scripts" (OuterVolumeSpecName: "scripts") pod "e41e549f-4050-4d3e-bad3-71e3ade56f86" (UID: "e41e549f-4050-4d3e-bad3-71e3ade56f86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.463115 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e41e549f-4050-4d3e-bad3-71e3ade56f86-kube-api-access-9t4l7" (OuterVolumeSpecName: "kube-api-access-9t4l7") pod "e41e549f-4050-4d3e-bad3-71e3ade56f86" (UID: "e41e549f-4050-4d3e-bad3-71e3ade56f86"). InnerVolumeSpecName "kube-api-access-9t4l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.557316 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e41e549f-4050-4d3e-bad3-71e3ade56f86-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.557347 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t4l7\" (UniqueName: \"kubernetes.io/projected/e41e549f-4050-4d3e-bad3-71e3ade56f86-kube-api-access-9t4l7\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.836234 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9grv6" event={"ID":"36d75e4f-e760-41d2-b7d8-b9baffb60bcf","Type":"ContainerStarted","Data":"1564a2ea10c090dbdeccf918d54f3a7fa0082ac5ca522861ca861a839b46bf97"} Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.839202 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z5csg-config-m6z5s" event={"ID":"e41e549f-4050-4d3e-bad3-71e3ade56f86","Type":"ContainerDied","Data":"fa58a9366e0a998c0b244b633ba8fe741812c12df39c77ddb6b202bf854eda81"} Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.839248 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa58a9366e0a998c0b244b633ba8fe741812c12df39c77ddb6b202bf854eda81" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.839305 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z5csg-config-m6z5s" Nov 22 08:24:00 crc kubenswrapper[4735]: I1122 08:24:00.862387 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-9grv6" podStartSLOduration=2.593393512 podStartE2EDuration="22.862361934s" podCreationTimestamp="2025-11-22 08:23:38 +0000 UTC" firstStartedPulling="2025-11-22 08:23:39.880140574 +0000 UTC m=+1241.484479179" lastFinishedPulling="2025-11-22 08:24:00.149108996 +0000 UTC m=+1261.753447601" observedRunningTime="2025-11-22 08:24:00.850429401 +0000 UTC m=+1262.454768006" watchObservedRunningTime="2025-11-22 08:24:00.862361934 +0000 UTC m=+1262.466700549" Nov 22 08:24:01 crc kubenswrapper[4735]: I1122 08:24:01.351878 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-z5csg-config-m6z5s"] Nov 22 08:24:01 crc kubenswrapper[4735]: I1122 08:24:01.362083 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-z5csg-config-m6z5s"] Nov 22 08:24:01 crc kubenswrapper[4735]: I1122 08:24:01.857842 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"8ba02a02c50009af5620c024e39362551f80ccd6f5bf8264aeaeb20242922bb6"} Nov 22 08:24:01 crc kubenswrapper[4735]: I1122 08:24:01.858405 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"2e4757baa78a6f84be1e27748b8943841f96e931a6a152935b515eedbe0f7906"} Nov 22 08:24:01 crc kubenswrapper[4735]: I1122 08:24:01.858539 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"9ceb9a9aa979ccde9d0e54507f36b7b0ed7c4bf71dae859fd797da3e8e0b99c2"} Nov 22 08:24:01 crc kubenswrapper[4735]: I1122 08:24:01.858622 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"bba39b9c23c3e51942c0c40caba12caf5488458db2aec67e68722120e7bae458"} Nov 22 08:24:03 crc kubenswrapper[4735]: I1122 08:24:03.283239 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e41e549f-4050-4d3e-bad3-71e3ade56f86" path="/var/lib/kubelet/pods/e41e549f-4050-4d3e-bad3-71e3ade56f86/volumes" Nov 22 08:24:03 crc kubenswrapper[4735]: I1122 08:24:03.893576 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"db5154287bd74b3f310d1d0c473e8d47a83ac650ee57a7aabc736571f88f5d6e"} Nov 22 08:24:03 crc kubenswrapper[4735]: I1122 08:24:03.893915 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"55b0c1e6351e34ff031f8fd6db859cb1e9ca546ebe5e5626d5de69122422c291"} Nov 22 08:24:03 crc kubenswrapper[4735]: I1122 08:24:03.893931 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"fce8e54b5c5adf43ac6ba6483e8298d142e2feede7968a279a91fa87718513fa"} Nov 22 08:24:03 crc kubenswrapper[4735]: I1122 08:24:03.893943 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"bd0305b7308d77c93a6bf478faaad175357c9e896346280caba16a42994419f9"} Nov 22 08:24:04 crc kubenswrapper[4735]: I1122 08:24:04.903394 4735 generic.go:334] "Generic (PLEG): container finished" podID="36d75e4f-e760-41d2-b7d8-b9baffb60bcf" containerID="1564a2ea10c090dbdeccf918d54f3a7fa0082ac5ca522861ca861a839b46bf97" exitCode=0 Nov 22 08:24:04 crc kubenswrapper[4735]: I1122 08:24:04.903481 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9grv6" event={"ID":"36d75e4f-e760-41d2-b7d8-b9baffb60bcf","Type":"ContainerDied","Data":"1564a2ea10c090dbdeccf918d54f3a7fa0082ac5ca522861ca861a839b46bf97"} Nov 22 08:24:04 crc kubenswrapper[4735]: I1122 08:24:04.909208 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"b15ff75a722b5a480a152bbf092a0a57bd68764939a48f425690b2894044898e"} Nov 22 08:24:04 crc kubenswrapper[4735]: I1122 08:24:04.909255 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"e9e985415c1c435f884e213f47596ea529bd24827b36c4b150d9504afcd0d7cc"} Nov 22 08:24:04 crc kubenswrapper[4735]: I1122 08:24:04.909269 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49e1ddee-0d04-4d19-9c87-e438e5e70e70","Type":"ContainerStarted","Data":"25f45fe71daeda491411ff0c117340e3f8cfe660d3755295eedd389fb29540a8"} Nov 22 08:24:04 crc kubenswrapper[4735]: I1122 08:24:04.965563 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.652032997 podStartE2EDuration="1m2.965546371s" podCreationTimestamp="2025-11-22 08:23:02 +0000 UTC" firstStartedPulling="2025-11-22 08:23:36.658330789 +0000 UTC m=+1238.262669394" lastFinishedPulling="2025-11-22 08:24:02.971844163 +0000 UTC m=+1264.576182768" observedRunningTime="2025-11-22 08:24:04.963249157 +0000 UTC m=+1266.567587762" watchObservedRunningTime="2025-11-22 08:24:04.965546371 +0000 UTC m=+1266.569884966" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.261422 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-t4dj8"] Nov 22 08:24:05 crc kubenswrapper[4735]: E1122 08:24:05.262060 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a566e05-aacd-44f9-b3dd-231549b9a967" containerName="mariadb-database-create" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262087 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a566e05-aacd-44f9-b3dd-231549b9a967" containerName="mariadb-database-create" Nov 22 08:24:05 crc kubenswrapper[4735]: E1122 08:24:05.262118 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="422ffe22-0786-4688-bd88-9e28481c59ff" containerName="mariadb-database-create" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262128 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="422ffe22-0786-4688-bd88-9e28481c59ff" containerName="mariadb-database-create" Nov 22 08:24:05 crc kubenswrapper[4735]: E1122 08:24:05.262164 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7" containerName="mariadb-account-create" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262173 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7" containerName="mariadb-account-create" Nov 22 08:24:05 crc kubenswrapper[4735]: E1122 08:24:05.262192 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6331445-03fa-4581-a892-974fdd4e53fa" containerName="mariadb-account-create" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262204 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6331445-03fa-4581-a892-974fdd4e53fa" containerName="mariadb-account-create" Nov 22 08:24:05 crc kubenswrapper[4735]: E1122 08:24:05.262218 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e41e549f-4050-4d3e-bad3-71e3ade56f86" containerName="ovn-config" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262226 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e41e549f-4050-4d3e-bad3-71e3ade56f86" containerName="ovn-config" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262480 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7" containerName="mariadb-account-create" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262513 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6331445-03fa-4581-a892-974fdd4e53fa" containerName="mariadb-account-create" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262528 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e41e549f-4050-4d3e-bad3-71e3ade56f86" containerName="ovn-config" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262601 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="422ffe22-0786-4688-bd88-9e28481c59ff" containerName="mariadb-database-create" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.262618 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a566e05-aacd-44f9-b3dd-231549b9a967" containerName="mariadb-database-create" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.264150 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.266619 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.285384 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-t4dj8"] Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.375073 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.375478 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-svc\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.375650 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.375833 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-config\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.375859 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl6xh\" (UniqueName: \"kubernetes.io/projected/4b8959b6-882d-4ada-bce5-b204536a16a2-kube-api-access-pl6xh\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.375900 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.477848 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.478049 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-config\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.478074 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl6xh\" (UniqueName: \"kubernetes.io/projected/4b8959b6-882d-4ada-bce5-b204536a16a2-kube-api-access-pl6xh\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.478115 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.478185 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.478229 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-svc\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.479202 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-svc\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.479835 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.480039 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.480514 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-config\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.480692 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.502245 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl6xh\" (UniqueName: \"kubernetes.io/projected/4b8959b6-882d-4ada-bce5-b204536a16a2-kube-api-access-pl6xh\") pod \"dnsmasq-dns-764c5664d7-t4dj8\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:05 crc kubenswrapper[4735]: I1122 08:24:05.591612 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.120897 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-t4dj8"] Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.244648 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9grv6" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.401862 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cprwc\" (UniqueName: \"kubernetes.io/projected/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-kube-api-access-cprwc\") pod \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.402967 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-config-data\") pod \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.403138 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-combined-ca-bundle\") pod \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\" (UID: \"36d75e4f-e760-41d2-b7d8-b9baffb60bcf\") " Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.409265 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-kube-api-access-cprwc" (OuterVolumeSpecName: "kube-api-access-cprwc") pod "36d75e4f-e760-41d2-b7d8-b9baffb60bcf" (UID: "36d75e4f-e760-41d2-b7d8-b9baffb60bcf"). InnerVolumeSpecName "kube-api-access-cprwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.436630 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36d75e4f-e760-41d2-b7d8-b9baffb60bcf" (UID: "36d75e4f-e760-41d2-b7d8-b9baffb60bcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.464009 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-config-data" (OuterVolumeSpecName: "config-data") pod "36d75e4f-e760-41d2-b7d8-b9baffb60bcf" (UID: "36d75e4f-e760-41d2-b7d8-b9baffb60bcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.506998 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cprwc\" (UniqueName: \"kubernetes.io/projected/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-kube-api-access-cprwc\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.507082 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.507105 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d75e4f-e760-41d2-b7d8-b9baffb60bcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.929012 4735 generic.go:334] "Generic (PLEG): container finished" podID="4b8959b6-882d-4ada-bce5-b204536a16a2" containerID="348156957a5b1fee9c882df5966b47e562545ff35846ef74e2929e920c2c4add" exitCode=0 Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.929091 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" event={"ID":"4b8959b6-882d-4ada-bce5-b204536a16a2","Type":"ContainerDied","Data":"348156957a5b1fee9c882df5966b47e562545ff35846ef74e2929e920c2c4add"} Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.929118 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" event={"ID":"4b8959b6-882d-4ada-bce5-b204536a16a2","Type":"ContainerStarted","Data":"9ca770ca0fc5a4b69bdfc3fdbebcd20c448702284b4a5000c9e0b50f3f7c22b6"} Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.931339 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9grv6" Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.931242 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9grv6" event={"ID":"36d75e4f-e760-41d2-b7d8-b9baffb60bcf","Type":"ContainerDied","Data":"9766c3b78da459609da1c207f9a2ccf206cb6007b7b78dd2e35bf5a500ca65db"} Nov 22 08:24:06 crc kubenswrapper[4735]: I1122 08:24:06.932674 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9766c3b78da459609da1c207f9a2ccf206cb6007b7b78dd2e35bf5a500ca65db" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.213241 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-t4dj8"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.228892 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mdc9q"] Nov 22 08:24:07 crc kubenswrapper[4735]: E1122 08:24:07.230000 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d75e4f-e760-41d2-b7d8-b9baffb60bcf" containerName="keystone-db-sync" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.230023 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d75e4f-e760-41d2-b7d8-b9baffb60bcf" containerName="keystone-db-sync" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.230244 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d75e4f-e760-41d2-b7d8-b9baffb60bcf" containerName="keystone-db-sync" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.230986 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.232738 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s4xtc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.236802 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.237020 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.237135 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.237316 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.255666 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mdc9q"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.313341 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-b62zz"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.315724 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.327489 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-fernet-keys\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.327538 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-config-data\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.327632 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fjcv\" (UniqueName: \"kubernetes.io/projected/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-kube-api-access-8fjcv\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.327673 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-scripts\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.327691 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-credential-keys\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.327750 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-combined-ca-bundle\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.328181 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-b62zz"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.384844 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-kvl85"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.387189 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.392568 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-qpwzp" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.392646 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429242 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429292 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-kvl85"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429332 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-combined-ca-bundle\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429439 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429490 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-svc\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429543 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-fernet-keys\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429573 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-config-data\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429646 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429680 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-config\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429737 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqb47\" (UniqueName: \"kubernetes.io/projected/8d84a8d4-188b-495e-8f2f-5595475f22d5-kube-api-access-wqb47\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429862 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fjcv\" (UniqueName: \"kubernetes.io/projected/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-kube-api-access-8fjcv\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429953 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-scripts\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.429984 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-credential-keys\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.438090 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-scripts\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.438958 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-credential-keys\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.440097 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-config-data\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.439313 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-fernet-keys\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.440958 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-combined-ca-bundle\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.448338 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fjcv\" (UniqueName: \"kubernetes.io/projected/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-kube-api-access-8fjcv\") pod \"keystone-bootstrap-mdc9q\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.528333 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-qsp4r"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.530120 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.532303 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblfx\" (UniqueName: \"kubernetes.io/projected/73149717-cef1-4092-b8af-5cceb0cd4830-kube-api-access-nblfx\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.532344 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.532369 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-config-data\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.532405 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.532423 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-svc\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.532811 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-combined-ca-bundle\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.532889 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.532960 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-config\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.533037 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqb47\" (UniqueName: \"kubernetes.io/projected/8d84a8d4-188b-495e-8f2f-5595475f22d5-kube-api-access-wqb47\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.533727 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.534382 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.534401 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.534698 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-config\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.534982 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.538634 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-svc\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.544882 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-chn6w" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.545013 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.557844 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.571547 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqb47\" (UniqueName: \"kubernetes.io/projected/8d84a8d4-188b-495e-8f2f-5595475f22d5-kube-api-access-wqb47\") pod \"dnsmasq-dns-5959f8865f-b62zz\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.604819 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-qsp4r"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.627649 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-qptkr"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.633675 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.634977 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-combined-ca-bundle\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.635143 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-scripts\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.635187 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqjnv\" (UniqueName: \"kubernetes.io/projected/c9f75ef9-02e1-4af0-b532-2e744d02d842-kube-api-access-tqjnv\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.635226 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-config-data\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.635258 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblfx\" (UniqueName: \"kubernetes.io/projected/73149717-cef1-4092-b8af-5cceb0cd4830-kube-api-access-nblfx\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.635296 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-config-data\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.635334 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c9f75ef9-02e1-4af0-b532-2e744d02d842-etc-machine-id\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.635369 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-db-sync-config-data\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.635399 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-combined-ca-bundle\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.638288 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.638558 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8sbsw" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.643176 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.659188 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-config-data\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.665224 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-combined-ca-bundle\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.679697 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qptkr"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.680273 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.683424 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblfx\" (UniqueName: \"kubernetes.io/projected/73149717-cef1-4092-b8af-5cceb0cd4830-kube-api-access-nblfx\") pod \"heat-db-sync-kvl85\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.713854 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.739361 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c9f75ef9-02e1-4af0-b532-2e744d02d842-etc-machine-id\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.739403 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-db-sync-config-data\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.739425 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-config\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.741182 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c9f75ef9-02e1-4af0-b532-2e744d02d842-etc-machine-id\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.741248 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-combined-ca-bundle\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.741726 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-combined-ca-bundle\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.741841 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7tt4\" (UniqueName: \"kubernetes.io/projected/f525dad0-8aeb-4928-b87f-561f2c3f3228-kube-api-access-m7tt4\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.741884 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-scripts\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.741902 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqjnv\" (UniqueName: \"kubernetes.io/projected/c9f75ef9-02e1-4af0-b532-2e744d02d842-kube-api-access-tqjnv\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.741928 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-config-data\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.746450 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nj6dc"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.754768 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-combined-ca-bundle\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.762927 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.764377 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-db-sync-config-data\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.767876 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-c9rr6" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.768168 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.768306 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.769223 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-scripts\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.772205 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-z64cd"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.774815 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.783850 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.783864 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqjnv\" (UniqueName: \"kubernetes.io/projected/c9f75ef9-02e1-4af0-b532-2e744d02d842-kube-api-access-tqjnv\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.785402 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mk4xm" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.813132 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-config-data\") pod \"cinder-db-sync-qsp4r\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.824073 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nj6dc"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.843395 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.845812 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a560c2a6-316e-4d1c-8b84-4a44aec7e772-logs\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.845879 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-combined-ca-bundle\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.845936 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbxk4\" (UniqueName: \"kubernetes.io/projected/e519a2d0-063e-4097-80f8-f4276208317a-kube-api-access-fbxk4\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.845991 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqbrl\" (UniqueName: \"kubernetes.io/projected/a560c2a6-316e-4d1c-8b84-4a44aec7e772-kube-api-access-tqbrl\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.846063 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7tt4\" (UniqueName: \"kubernetes.io/projected/f525dad0-8aeb-4928-b87f-561f2c3f3228-kube-api-access-m7tt4\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.846159 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-config-data\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.846194 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-db-sync-config-data\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.846220 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-combined-ca-bundle\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.846280 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-scripts\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.846316 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-config\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.846361 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-combined-ca-bundle\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.864887 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7tt4\" (UniqueName: \"kubernetes.io/projected/f525dad0-8aeb-4928-b87f-561f2c3f3228-kube-api-access-m7tt4\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.869790 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-combined-ca-bundle\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.875362 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-z64cd"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.878345 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-config\") pod \"neutron-db-sync-qptkr\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.915991 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-b62zz"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.955810 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-config-data\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.955863 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-db-sync-config-data\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.955892 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-combined-ca-bundle\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.955965 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-scripts\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.956123 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a560c2a6-316e-4d1c-8b84-4a44aec7e772-logs\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.956174 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-combined-ca-bundle\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.956214 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbxk4\" (UniqueName: \"kubernetes.io/projected/e519a2d0-063e-4097-80f8-f4276208317a-kube-api-access-fbxk4\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.956289 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqbrl\" (UniqueName: \"kubernetes.io/projected/a560c2a6-316e-4d1c-8b84-4a44aec7e772-kube-api-access-tqbrl\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.958295 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a560c2a6-316e-4d1c-8b84-4a44aec7e772-logs\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.964046 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-7cmlv"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.966095 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.968329 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-combined-ca-bundle\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.980875 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.993068 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-scripts\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.993418 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-db-sync-config-data\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.993538 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" podUID="4b8959b6-882d-4ada-bce5-b204536a16a2" containerName="dnsmasq-dns" containerID="cri-o://e7cf03d5986fae3bfbad4db3f9f60e2e05d51d4e99ca71ecc60f0a00104622ea" gracePeriod=10 Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.993984 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-config-data\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:07 crc kubenswrapper[4735]: I1122 08:24:07.998395 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-combined-ca-bundle\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.000377 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.000389 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqbrl\" (UniqueName: \"kubernetes.io/projected/a560c2a6-316e-4d1c-8b84-4a44aec7e772-kube-api-access-tqbrl\") pod \"placement-db-sync-nj6dc\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.000405 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" event={"ID":"4b8959b6-882d-4ada-bce5-b204536a16a2","Type":"ContainerStarted","Data":"e7cf03d5986fae3bfbad4db3f9f60e2e05d51d4e99ca71ecc60f0a00104622ea"} Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.000509 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-7cmlv"] Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.000690 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.002534 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.005433 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.009633 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbxk4\" (UniqueName: \"kubernetes.io/projected/e519a2d0-063e-4097-80f8-f4276208317a-kube-api-access-fbxk4\") pod \"barbican-db-sync-z64cd\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.019390 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061549 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061599 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061632 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061666 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061697 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smh5b\" (UniqueName: \"kubernetes.io/projected/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-kube-api-access-smh5b\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061714 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-config-data\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061749 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-run-httpd\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061785 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061810 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-log-httpd\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061859 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-config\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061879 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-scripts\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061902 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tx42\" (UniqueName: \"kubernetes.io/projected/e6c6422b-0bb3-4c77-82dc-540f707ab181-kube-api-access-2tx42\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.061924 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.102496 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" podStartSLOduration=3.102442195 podStartE2EDuration="3.102442195s" podCreationTimestamp="2025-11-22 08:24:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:08.022024768 +0000 UTC m=+1269.626363363" watchObservedRunningTime="2025-11-22 08:24:08.102442195 +0000 UTC m=+1269.706780800" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.155010 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.164682 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-log-httpd\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.164765 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-config\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.164791 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-scripts\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.164824 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tx42\" (UniqueName: \"kubernetes.io/projected/e6c6422b-0bb3-4c77-82dc-540f707ab181-kube-api-access-2tx42\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.164856 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.165250 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.165289 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.165326 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.165366 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.165391 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smh5b\" (UniqueName: \"kubernetes.io/projected/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-kube-api-access-smh5b\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.165409 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-config-data\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.165448 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-run-httpd\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.166341 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.166858 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-log-httpd\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.167072 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.167796 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.168508 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-config\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.172148 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-run-httpd\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.176451 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.177591 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.177983 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.182723 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-config-data\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.184048 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.184246 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-scripts\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.184369 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.184554 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z64cd" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.193298 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smh5b\" (UniqueName: \"kubernetes.io/projected/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-kube-api-access-smh5b\") pod \"ceilometer-0\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.203107 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tx42\" (UniqueName: \"kubernetes.io/projected/e6c6422b-0bb3-4c77-82dc-540f707ab181-kube-api-access-2tx42\") pod \"dnsmasq-dns-58dd9ff6bc-7cmlv\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.305113 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.325328 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:24:08 crc kubenswrapper[4735]: E1122 08:24:08.510951 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b8959b6_882d_4ada_bce5_b204536a16a2.slice/crio-e7cf03d5986fae3bfbad4db3f9f60e2e05d51d4e99ca71ecc60f0a00104622ea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b8959b6_882d_4ada_bce5_b204536a16a2.slice/crio-conmon-e7cf03d5986fae3bfbad4db3f9f60e2e05d51d4e99ca71ecc60f0a00104622ea.scope\": RecentStats: unable to find data in memory cache]" Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.787842 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-b62zz"] Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.799106 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mdc9q"] Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.807369 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-kvl85"] Nov 22 08:24:08 crc kubenswrapper[4735]: I1122 08:24:08.849069 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-qsp4r"] Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.022405 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qsp4r" event={"ID":"c9f75ef9-02e1-4af0-b532-2e744d02d842","Type":"ContainerStarted","Data":"3a2c6fa0b1cc36f0dd3c29b222c23070d1b5e5870b62e5953d1fb13555a62e62"} Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.037478 4735 generic.go:334] "Generic (PLEG): container finished" podID="4b8959b6-882d-4ada-bce5-b204536a16a2" containerID="e7cf03d5986fae3bfbad4db3f9f60e2e05d51d4e99ca71ecc60f0a00104622ea" exitCode=0 Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.037563 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" event={"ID":"4b8959b6-882d-4ada-bce5-b204536a16a2","Type":"ContainerDied","Data":"e7cf03d5986fae3bfbad4db3f9f60e2e05d51d4e99ca71ecc60f0a00104622ea"} Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.037592 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" event={"ID":"4b8959b6-882d-4ada-bce5-b204536a16a2","Type":"ContainerDied","Data":"9ca770ca0fc5a4b69bdfc3fdbebcd20c448702284b4a5000c9e0b50f3f7c22b6"} Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.037606 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ca770ca0fc5a4b69bdfc3fdbebcd20c448702284b4a5000c9e0b50f3f7c22b6" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.038938 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-b62zz" event={"ID":"8d84a8d4-188b-495e-8f2f-5595475f22d5","Type":"ContainerStarted","Data":"f4e0df265ff93dc59f65ad6aa499254050a00242c20c5248cf9a4afc809c26df"} Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.046692 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-m9x5r" event={"ID":"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3","Type":"ContainerStarted","Data":"82367da6640e3280143e869e40240f2b5acad22fd6e4644929227988c5465988"} Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.058617 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-kvl85" event={"ID":"73149717-cef1-4092-b8af-5cceb0cd4830","Type":"ContainerStarted","Data":"d15cfb6d7b46b8afe34c869772320d9c7b3670a66523f1f5da90ea2436f17c37"} Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.068685 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mdc9q" event={"ID":"ced0dd80-76f3-4e26-9ed7-f7d516f028f3","Type":"ContainerStarted","Data":"96ce3368e46be7c3895275f54bb4bf94f2d10f57200cf468673b8e56e44f5498"} Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.072691 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-m9x5r" podStartSLOduration=3.566061795 podStartE2EDuration="43.072663478s" podCreationTimestamp="2025-11-22 08:23:26 +0000 UTC" firstStartedPulling="2025-11-22 08:23:27.485387988 +0000 UTC m=+1229.089726613" lastFinishedPulling="2025-11-22 08:24:06.991989691 +0000 UTC m=+1268.596328296" observedRunningTime="2025-11-22 08:24:09.072279408 +0000 UTC m=+1270.676618023" watchObservedRunningTime="2025-11-22 08:24:09.072663478 +0000 UTC m=+1270.677002083" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.080853 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.222340 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-config\") pod \"4b8959b6-882d-4ada-bce5-b204536a16a2\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.222487 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-nb\") pod \"4b8959b6-882d-4ada-bce5-b204536a16a2\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.222659 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-sb\") pod \"4b8959b6-882d-4ada-bce5-b204536a16a2\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.222826 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-swift-storage-0\") pod \"4b8959b6-882d-4ada-bce5-b204536a16a2\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.222885 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-svc\") pod \"4b8959b6-882d-4ada-bce5-b204536a16a2\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.222915 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl6xh\" (UniqueName: \"kubernetes.io/projected/4b8959b6-882d-4ada-bce5-b204536a16a2-kube-api-access-pl6xh\") pod \"4b8959b6-882d-4ada-bce5-b204536a16a2\" (UID: \"4b8959b6-882d-4ada-bce5-b204536a16a2\") " Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.228388 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b8959b6-882d-4ada-bce5-b204536a16a2-kube-api-access-pl6xh" (OuterVolumeSpecName: "kube-api-access-pl6xh") pod "4b8959b6-882d-4ada-bce5-b204536a16a2" (UID: "4b8959b6-882d-4ada-bce5-b204536a16a2"). InnerVolumeSpecName "kube-api-access-pl6xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.234312 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-z64cd"] Nov 22 08:24:09 crc kubenswrapper[4735]: W1122 08:24:09.259023 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode519a2d0_063e_4097_80f8_f4276208317a.slice/crio-9a1ef4f014665f0acfe34cb65044872baa410f461756acfcd9cd9fbc6eb66d60 WatchSource:0}: Error finding container 9a1ef4f014665f0acfe34cb65044872baa410f461756acfcd9cd9fbc6eb66d60: Status 404 returned error can't find the container with id 9a1ef4f014665f0acfe34cb65044872baa410f461756acfcd9cd9fbc6eb66d60 Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.324853 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qptkr"] Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.326504 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl6xh\" (UniqueName: \"kubernetes.io/projected/4b8959b6-882d-4ada-bce5-b204536a16a2-kube-api-access-pl6xh\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.345452 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nj6dc"] Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.369681 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-7cmlv"] Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.382502 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.753736 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-config" (OuterVolumeSpecName: "config") pod "4b8959b6-882d-4ada-bce5-b204536a16a2" (UID: "4b8959b6-882d-4ada-bce5-b204536a16a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.789322 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4b8959b6-882d-4ada-bce5-b204536a16a2" (UID: "4b8959b6-882d-4ada-bce5-b204536a16a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.789850 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4b8959b6-882d-4ada-bce5-b204536a16a2" (UID: "4b8959b6-882d-4ada-bce5-b204536a16a2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.791500 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4b8959b6-882d-4ada-bce5-b204536a16a2" (UID: "4b8959b6-882d-4ada-bce5-b204536a16a2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.807744 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4b8959b6-882d-4ada-bce5-b204536a16a2" (UID: "4b8959b6-882d-4ada-bce5-b204536a16a2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.859958 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.859996 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.860011 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.860022 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.860033 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b8959b6-882d-4ada-bce5-b204536a16a2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:09 crc kubenswrapper[4735]: I1122 08:24:09.910999 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.086030 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" event={"ID":"e6c6422b-0bb3-4c77-82dc-540f707ab181","Type":"ContainerStarted","Data":"964d656c9dbff9fd6ee385c89b07045eb83deefe3f6fd96ad7d00b1c1620d0e7"} Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.088331 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qptkr" event={"ID":"f525dad0-8aeb-4928-b87f-561f2c3f3228","Type":"ContainerStarted","Data":"0b9ce23a551ece580c6f82eadd857d34497ca3c41f1b99f1f06eee7cfbad15eb"} Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.125998 4735 generic.go:334] "Generic (PLEG): container finished" podID="8d84a8d4-188b-495e-8f2f-5595475f22d5" containerID="b81294bc1d3027a9b19a08d76b97b579b8d44243861e80f7e72c1a35af2d9886" exitCode=0 Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.126059 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-b62zz" event={"ID":"8d84a8d4-188b-495e-8f2f-5595475f22d5","Type":"ContainerDied","Data":"b81294bc1d3027a9b19a08d76b97b579b8d44243861e80f7e72c1a35af2d9886"} Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.140446 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nj6dc" event={"ID":"a560c2a6-316e-4d1c-8b84-4a44aec7e772","Type":"ContainerStarted","Data":"e9bc50eb9eb0e54d229d4e7b1f9e9d1840500a3aeedfd5da01397ba1879112d5"} Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.146759 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-qptkr" podStartSLOduration=3.146738241 podStartE2EDuration="3.146738241s" podCreationTimestamp="2025-11-22 08:24:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:10.130824558 +0000 UTC m=+1271.735163163" watchObservedRunningTime="2025-11-22 08:24:10.146738241 +0000 UTC m=+1271.751076846" Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.172513 4735 generic.go:334] "Generic (PLEG): container finished" podID="f1d9efee-4efc-460c-ba99-36df95d96a48" containerID="264d532f6ae9cd12781153d0dcb50e74bd0c3beba046296f86e4bbb734e78963" exitCode=0 Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.172596 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f1d9efee-4efc-460c-ba99-36df95d96a48","Type":"ContainerDied","Data":"264d532f6ae9cd12781153d0dcb50e74bd0c3beba046296f86e4bbb734e78963"} Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.195789 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z64cd" event={"ID":"e519a2d0-063e-4097-80f8-f4276208317a","Type":"ContainerStarted","Data":"9a1ef4f014665f0acfe34cb65044872baa410f461756acfcd9cd9fbc6eb66d60"} Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.218082 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mdc9q" event={"ID":"ced0dd80-76f3-4e26-9ed7-f7d516f028f3","Type":"ContainerStarted","Data":"ecbdc6d9ee393531116e462cba48b131422b3d535366727877f5f339eae70cf2"} Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.226566 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerStarted","Data":"26c2b5f297b401d3866e16b2cb1be602e30f09f9d824bbeb0beab395d956a791"} Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.226613 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-t4dj8" Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.413325 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mdc9q" podStartSLOduration=3.413301406 podStartE2EDuration="3.413301406s" podCreationTimestamp="2025-11-22 08:24:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:10.245620321 +0000 UTC m=+1271.849958926" watchObservedRunningTime="2025-11-22 08:24:10.413301406 +0000 UTC m=+1272.017640011" Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.432493 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-t4dj8"] Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.477417 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-t4dj8"] Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.831185 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.913193 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqb47\" (UniqueName: \"kubernetes.io/projected/8d84a8d4-188b-495e-8f2f-5595475f22d5-kube-api-access-wqb47\") pod \"8d84a8d4-188b-495e-8f2f-5595475f22d5\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.913264 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-config\") pod \"8d84a8d4-188b-495e-8f2f-5595475f22d5\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.913410 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-svc\") pod \"8d84a8d4-188b-495e-8f2f-5595475f22d5\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.913490 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-swift-storage-0\") pod \"8d84a8d4-188b-495e-8f2f-5595475f22d5\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.917539 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-nb\") pod \"8d84a8d4-188b-495e-8f2f-5595475f22d5\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.917784 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-sb\") pod \"8d84a8d4-188b-495e-8f2f-5595475f22d5\" (UID: \"8d84a8d4-188b-495e-8f2f-5595475f22d5\") " Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.923118 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d84a8d4-188b-495e-8f2f-5595475f22d5-kube-api-access-wqb47" (OuterVolumeSpecName: "kube-api-access-wqb47") pod "8d84a8d4-188b-495e-8f2f-5595475f22d5" (UID: "8d84a8d4-188b-495e-8f2f-5595475f22d5"). InnerVolumeSpecName "kube-api-access-wqb47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.948582 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8d84a8d4-188b-495e-8f2f-5595475f22d5" (UID: "8d84a8d4-188b-495e-8f2f-5595475f22d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.949794 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8d84a8d4-188b-495e-8f2f-5595475f22d5" (UID: "8d84a8d4-188b-495e-8f2f-5595475f22d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:10 crc kubenswrapper[4735]: I1122 08:24:10.957058 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-config" (OuterVolumeSpecName: "config") pod "8d84a8d4-188b-495e-8f2f-5595475f22d5" (UID: "8d84a8d4-188b-495e-8f2f-5595475f22d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.021750 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqb47\" (UniqueName: \"kubernetes.io/projected/8d84a8d4-188b-495e-8f2f-5595475f22d5-kube-api-access-wqb47\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.021796 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.021812 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.021823 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.064475 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8d84a8d4-188b-495e-8f2f-5595475f22d5" (UID: "8d84a8d4-188b-495e-8f2f-5595475f22d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.070557 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8d84a8d4-188b-495e-8f2f-5595475f22d5" (UID: "8d84a8d4-188b-495e-8f2f-5595475f22d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.124079 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.124109 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d84a8d4-188b-495e-8f2f-5595475f22d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.262358 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-b62zz" event={"ID":"8d84a8d4-188b-495e-8f2f-5595475f22d5","Type":"ContainerDied","Data":"f4e0df265ff93dc59f65ad6aa499254050a00242c20c5248cf9a4afc809c26df"} Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.262421 4735 scope.go:117] "RemoveContainer" containerID="b81294bc1d3027a9b19a08d76b97b579b8d44243861e80f7e72c1a35af2d9886" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.262622 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-b62zz" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.284940 4735 generic.go:334] "Generic (PLEG): container finished" podID="e6c6422b-0bb3-4c77-82dc-540f707ab181" containerID="90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b" exitCode=0 Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.296515 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b8959b6-882d-4ada-bce5-b204536a16a2" path="/var/lib/kubelet/pods/4b8959b6-882d-4ada-bce5-b204536a16a2/volumes" Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.297401 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f1d9efee-4efc-460c-ba99-36df95d96a48","Type":"ContainerStarted","Data":"456fb80ced1b1271a4aa89ca57356993ea80bb2c276484a7d4a362a982853b5c"} Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.297429 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" event={"ID":"e6c6422b-0bb3-4c77-82dc-540f707ab181","Type":"ContainerDied","Data":"90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b"} Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.317521 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qptkr" event={"ID":"f525dad0-8aeb-4928-b87f-561f2c3f3228","Type":"ContainerStarted","Data":"427d0f580305583837f5847a430716a462eec9578979f975200112fc0c9a452c"} Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.361517 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-b62zz"] Nov 22 08:24:11 crc kubenswrapper[4735]: I1122 08:24:11.370821 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-b62zz"] Nov 22 08:24:12 crc kubenswrapper[4735]: I1122 08:24:12.337939 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" event={"ID":"e6c6422b-0bb3-4c77-82dc-540f707ab181","Type":"ContainerStarted","Data":"ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80"} Nov 22 08:24:12 crc kubenswrapper[4735]: I1122 08:24:12.338330 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:12 crc kubenswrapper[4735]: I1122 08:24:12.365180 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" podStartSLOduration=5.365158592 podStartE2EDuration="5.365158592s" podCreationTimestamp="2025-11-22 08:24:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:12.355815821 +0000 UTC m=+1273.960154436" watchObservedRunningTime="2025-11-22 08:24:12.365158592 +0000 UTC m=+1273.969497197" Nov 22 08:24:13 crc kubenswrapper[4735]: I1122 08:24:13.282486 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d84a8d4-188b-495e-8f2f-5595475f22d5" path="/var/lib/kubelet/pods/8d84a8d4-188b-495e-8f2f-5595475f22d5/volumes" Nov 22 08:24:14 crc kubenswrapper[4735]: I1122 08:24:14.374265 4735 generic.go:334] "Generic (PLEG): container finished" podID="ced0dd80-76f3-4e26-9ed7-f7d516f028f3" containerID="ecbdc6d9ee393531116e462cba48b131422b3d535366727877f5f339eae70cf2" exitCode=0 Nov 22 08:24:14 crc kubenswrapper[4735]: I1122 08:24:14.374368 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mdc9q" event={"ID":"ced0dd80-76f3-4e26-9ed7-f7d516f028f3","Type":"ContainerDied","Data":"ecbdc6d9ee393531116e462cba48b131422b3d535366727877f5f339eae70cf2"} Nov 22 08:24:15 crc kubenswrapper[4735]: I1122 08:24:15.389481 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f1d9efee-4efc-460c-ba99-36df95d96a48","Type":"ContainerStarted","Data":"79b9afd77838c035e65f28559075820c1aada120b07b010eed7aedc924731d2b"} Nov 22 08:24:16 crc kubenswrapper[4735]: I1122 08:24:16.141106 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:24:16 crc kubenswrapper[4735]: I1122 08:24:16.141447 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:24:18 crc kubenswrapper[4735]: I1122 08:24:18.307407 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:18 crc kubenswrapper[4735]: I1122 08:24:18.359374 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nrccs"] Nov 22 08:24:18 crc kubenswrapper[4735]: I1122 08:24:18.359600 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-nrccs" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="dnsmasq-dns" containerID="cri-o://ced057972c34b33ab51b58b80d8b4d0fb8a52b95f5cf6bf5837b966ff7a17164" gracePeriod=10 Nov 22 08:24:19 crc kubenswrapper[4735]: I1122 08:24:19.440523 4735 generic.go:334] "Generic (PLEG): container finished" podID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerID="ced057972c34b33ab51b58b80d8b4d0fb8a52b95f5cf6bf5837b966ff7a17164" exitCode=0 Nov 22 08:24:19 crc kubenswrapper[4735]: I1122 08:24:19.440590 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nrccs" event={"ID":"3ee474f0-0a9f-4c80-bacb-6140ac9c571c","Type":"ContainerDied","Data":"ced057972c34b33ab51b58b80d8b4d0fb8a52b95f5cf6bf5837b966ff7a17164"} Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.327980 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.450281 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-combined-ca-bundle\") pod \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.450608 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-config-data\") pod \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.450659 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-fernet-keys\") pod \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.450715 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-credential-keys\") pod \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.450739 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fjcv\" (UniqueName: \"kubernetes.io/projected/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-kube-api-access-8fjcv\") pod \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.450830 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-scripts\") pod \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\" (UID: \"ced0dd80-76f3-4e26-9ed7-f7d516f028f3\") " Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.453284 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mdc9q" event={"ID":"ced0dd80-76f3-4e26-9ed7-f7d516f028f3","Type":"ContainerDied","Data":"96ce3368e46be7c3895275f54bb4bf94f2d10f57200cf468673b8e56e44f5498"} Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.453334 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96ce3368e46be7c3895275f54bb4bf94f2d10f57200cf468673b8e56e44f5498" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.453345 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mdc9q" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.457021 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ced0dd80-76f3-4e26-9ed7-f7d516f028f3" (UID: "ced0dd80-76f3-4e26-9ed7-f7d516f028f3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.457585 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ced0dd80-76f3-4e26-9ed7-f7d516f028f3" (UID: "ced0dd80-76f3-4e26-9ed7-f7d516f028f3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.457658 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-kube-api-access-8fjcv" (OuterVolumeSpecName: "kube-api-access-8fjcv") pod "ced0dd80-76f3-4e26-9ed7-f7d516f028f3" (UID: "ced0dd80-76f3-4e26-9ed7-f7d516f028f3"). InnerVolumeSpecName "kube-api-access-8fjcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.465682 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-scripts" (OuterVolumeSpecName: "scripts") pod "ced0dd80-76f3-4e26-9ed7-f7d516f028f3" (UID: "ced0dd80-76f3-4e26-9ed7-f7d516f028f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.466211 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-nrccs" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: connect: connection refused" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.483829 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ced0dd80-76f3-4e26-9ed7-f7d516f028f3" (UID: "ced0dd80-76f3-4e26-9ed7-f7d516f028f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.490970 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-config-data" (OuterVolumeSpecName: "config-data") pod "ced0dd80-76f3-4e26-9ed7-f7d516f028f3" (UID: "ced0dd80-76f3-4e26-9ed7-f7d516f028f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.553125 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.553154 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.553163 4735 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.553171 4735 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.553179 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fjcv\" (UniqueName: \"kubernetes.io/projected/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-kube-api-access-8fjcv\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:20 crc kubenswrapper[4735]: I1122 08:24:20.553189 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced0dd80-76f3-4e26-9ed7-f7d516f028f3-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.437588 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mdc9q"] Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.461896 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mdc9q"] Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.480267 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-t9c5x"] Nov 22 08:24:21 crc kubenswrapper[4735]: E1122 08:24:21.480743 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced0dd80-76f3-4e26-9ed7-f7d516f028f3" containerName="keystone-bootstrap" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.480762 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced0dd80-76f3-4e26-9ed7-f7d516f028f3" containerName="keystone-bootstrap" Nov 22 08:24:21 crc kubenswrapper[4735]: E1122 08:24:21.480779 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8959b6-882d-4ada-bce5-b204536a16a2" containerName="init" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.480787 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8959b6-882d-4ada-bce5-b204536a16a2" containerName="init" Nov 22 08:24:21 crc kubenswrapper[4735]: E1122 08:24:21.480801 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8959b6-882d-4ada-bce5-b204536a16a2" containerName="dnsmasq-dns" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.480807 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8959b6-882d-4ada-bce5-b204536a16a2" containerName="dnsmasq-dns" Nov 22 08:24:21 crc kubenswrapper[4735]: E1122 08:24:21.480830 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d84a8d4-188b-495e-8f2f-5595475f22d5" containerName="init" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.480836 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d84a8d4-188b-495e-8f2f-5595475f22d5" containerName="init" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.481033 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d84a8d4-188b-495e-8f2f-5595475f22d5" containerName="init" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.481212 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced0dd80-76f3-4e26-9ed7-f7d516f028f3" containerName="keystone-bootstrap" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.481236 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b8959b6-882d-4ada-bce5-b204536a16a2" containerName="dnsmasq-dns" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.482338 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.484888 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.489153 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t9c5x"] Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.489292 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.490552 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.490915 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s4xtc" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.491586 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.581144 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-credential-keys\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.581184 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghxxd\" (UniqueName: \"kubernetes.io/projected/a32d2023-853f-40f2-9322-155b32c82fdb-kube-api-access-ghxxd\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.581204 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-config-data\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.581694 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-fernet-keys\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.581737 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-scripts\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.581796 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-combined-ca-bundle\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.683595 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-fernet-keys\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.683654 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-scripts\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.683687 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-combined-ca-bundle\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.683773 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-credential-keys\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.683802 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghxxd\" (UniqueName: \"kubernetes.io/projected/a32d2023-853f-40f2-9322-155b32c82fdb-kube-api-access-ghxxd\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.683823 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-config-data\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.688531 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-credential-keys\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.689733 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-config-data\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.692933 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-fernet-keys\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.694203 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-scripts\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.712279 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-combined-ca-bundle\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.727520 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghxxd\" (UniqueName: \"kubernetes.io/projected/a32d2023-853f-40f2-9322-155b32c82fdb-kube-api-access-ghxxd\") pod \"keystone-bootstrap-t9c5x\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:21 crc kubenswrapper[4735]: I1122 08:24:21.800136 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:22 crc kubenswrapper[4735]: E1122 08:24:22.155965 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 22 08:24:22 crc kubenswrapper[4735]: E1122 08:24:22.156684 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tqbrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-nj6dc_openstack(a560c2a6-316e-4d1c-8b84-4a44aec7e772): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:24:22 crc kubenswrapper[4735]: E1122 08:24:22.158120 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-nj6dc" podUID="a560c2a6-316e-4d1c-8b84-4a44aec7e772" Nov 22 08:24:22 crc kubenswrapper[4735]: E1122 08:24:22.487652 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-nj6dc" podUID="a560c2a6-316e-4d1c-8b84-4a44aec7e772" Nov 22 08:24:23 crc kubenswrapper[4735]: I1122 08:24:23.275075 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced0dd80-76f3-4e26-9ed7-f7d516f028f3" path="/var/lib/kubelet/pods/ced0dd80-76f3-4e26-9ed7-f7d516f028f3/volumes" Nov 22 08:24:24 crc kubenswrapper[4735]: I1122 08:24:24.511968 4735 generic.go:334] "Generic (PLEG): container finished" podID="54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" containerID="82367da6640e3280143e869e40240f2b5acad22fd6e4644929227988c5465988" exitCode=0 Nov 22 08:24:24 crc kubenswrapper[4735]: I1122 08:24:24.512075 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-m9x5r" event={"ID":"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3","Type":"ContainerDied","Data":"82367da6640e3280143e869e40240f2b5acad22fd6e4644929227988c5465988"} Nov 22 08:24:30 crc kubenswrapper[4735]: I1122 08:24:30.466299 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-nrccs" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: i/o timeout" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.178889 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.187219 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-m9x5r" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295327 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-config\") pod \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295489 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-sb\") pod \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295543 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-combined-ca-bundle\") pod \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295571 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-nb\") pod \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295687 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-dns-svc\") pod \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295733 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twfqg\" (UniqueName: \"kubernetes.io/projected/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-kube-api-access-twfqg\") pod \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295765 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z75m5\" (UniqueName: \"kubernetes.io/projected/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-kube-api-access-z75m5\") pod \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\" (UID: \"3ee474f0-0a9f-4c80-bacb-6140ac9c571c\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295796 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-config-data\") pod \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.295892 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-db-sync-config-data\") pod \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\" (UID: \"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3\") " Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.302521 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" (UID: "54265b3d-7ff9-4b71-aab0-9b79ae0b08d3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.304651 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-kube-api-access-twfqg" (OuterVolumeSpecName: "kube-api-access-twfqg") pod "54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" (UID: "54265b3d-7ff9-4b71-aab0-9b79ae0b08d3"). InnerVolumeSpecName "kube-api-access-twfqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.305379 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-kube-api-access-z75m5" (OuterVolumeSpecName: "kube-api-access-z75m5") pod "3ee474f0-0a9f-4c80-bacb-6140ac9c571c" (UID: "3ee474f0-0a9f-4c80-bacb-6140ac9c571c"). InnerVolumeSpecName "kube-api-access-z75m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.330825 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" (UID: "54265b3d-7ff9-4b71-aab0-9b79ae0b08d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.352166 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-config" (OuterVolumeSpecName: "config") pod "3ee474f0-0a9f-4c80-bacb-6140ac9c571c" (UID: "3ee474f0-0a9f-4c80-bacb-6140ac9c571c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.353728 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ee474f0-0a9f-4c80-bacb-6140ac9c571c" (UID: "3ee474f0-0a9f-4c80-bacb-6140ac9c571c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.355129 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ee474f0-0a9f-4c80-bacb-6140ac9c571c" (UID: "3ee474f0-0a9f-4c80-bacb-6140ac9c571c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.356757 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ee474f0-0a9f-4c80-bacb-6140ac9c571c" (UID: "3ee474f0-0a9f-4c80-bacb-6140ac9c571c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.371129 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-config-data" (OuterVolumeSpecName: "config-data") pod "54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" (UID: "54265b3d-7ff9-4b71-aab0-9b79ae0b08d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399439 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399484 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twfqg\" (UniqueName: \"kubernetes.io/projected/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-kube-api-access-twfqg\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399496 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z75m5\" (UniqueName: \"kubernetes.io/projected/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-kube-api-access-z75m5\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399505 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399514 4735 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399523 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399531 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399539 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.399547 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ee474f0-0a9f-4c80-bacb-6140ac9c571c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.467671 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-nrccs" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: i/o timeout" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.467782 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.635355 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-m9x5r" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.635363 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-m9x5r" event={"ID":"54265b3d-7ff9-4b71-aab0-9b79ae0b08d3","Type":"ContainerDied","Data":"301efc5175d965eebd6c21df22788f260d29a3a03feb01be64d3974a9c0ff3a1"} Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.635494 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="301efc5175d965eebd6c21df22788f260d29a3a03feb01be64d3974a9c0ff3a1" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.638023 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nrccs" event={"ID":"3ee474f0-0a9f-4c80-bacb-6140ac9c571c","Type":"ContainerDied","Data":"9d184d168ea858104727f4f2818845eb756e8ac24f6563405e7c9699e8622904"} Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.638062 4735 scope.go:117] "RemoveContainer" containerID="ced057972c34b33ab51b58b80d8b4d0fb8a52b95f5cf6bf5837b966ff7a17164" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.638161 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-nrccs" Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.694000 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nrccs"] Nov 22 08:24:35 crc kubenswrapper[4735]: I1122 08:24:35.703191 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nrccs"] Nov 22 08:24:35 crc kubenswrapper[4735]: E1122 08:24:35.737734 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 22 08:24:35 crc kubenswrapper[4735]: E1122 08:24:35.737897 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fbxk4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-z64cd_openstack(e519a2d0-063e-4097-80f8-f4276208317a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:24:35 crc kubenswrapper[4735]: E1122 08:24:35.739119 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-z64cd" podUID="e519a2d0-063e-4097-80f8-f4276208317a" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.597017 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-f4gcs"] Nov 22 08:24:36 crc kubenswrapper[4735]: E1122 08:24:36.597655 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="dnsmasq-dns" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.597667 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="dnsmasq-dns" Nov 22 08:24:36 crc kubenswrapper[4735]: E1122 08:24:36.597686 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" containerName="glance-db-sync" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.597692 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" containerName="glance-db-sync" Nov 22 08:24:36 crc kubenswrapper[4735]: E1122 08:24:36.597700 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="init" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.597708 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="init" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.597912 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" containerName="dnsmasq-dns" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.597923 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" containerName="glance-db-sync" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.600054 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.613725 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-f4gcs"] Nov 22 08:24:36 crc kubenswrapper[4735]: E1122 08:24:36.658248 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-z64cd" podUID="e519a2d0-063e-4097-80f8-f4276208317a" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.733411 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-config\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.733633 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.733687 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.733724 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l4hb\" (UniqueName: \"kubernetes.io/projected/bd32522c-f1ea-4f33-9e30-7c574e3b427d-kube-api-access-4l4hb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.733786 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.733813 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.835679 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.835766 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.835807 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l4hb\" (UniqueName: \"kubernetes.io/projected/bd32522c-f1ea-4f33-9e30-7c574e3b427d-kube-api-access-4l4hb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.835875 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.835909 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.835971 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-config\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.836598 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.836599 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.837066 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-config\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.837173 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.837254 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.860282 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l4hb\" (UniqueName: \"kubernetes.io/projected/bd32522c-f1ea-4f33-9e30-7c574e3b427d-kube-api-access-4l4hb\") pod \"dnsmasq-dns-785d8bcb8c-f4gcs\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:36 crc kubenswrapper[4735]: I1122 08:24:36.933594 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:37 crc kubenswrapper[4735]: E1122 08:24:37.139891 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 22 08:24:37 crc kubenswrapper[4735]: E1122 08:24:37.140405 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tqjnv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-qsp4r_openstack(c9f75ef9-02e1-4af0-b532-2e744d02d842): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:24:37 crc kubenswrapper[4735]: E1122 08:24:37.141979 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-qsp4r" podUID="c9f75ef9-02e1-4af0-b532-2e744d02d842" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.142605 4735 scope.go:117] "RemoveContainer" containerID="ec0ad732d773a8ecbde959d4860287f82f5db5c8095a3fa21e3a39061b5a6023" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.318568 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ee474f0-0a9f-4c80-bacb-6140ac9c571c" path="/var/lib/kubelet/pods/3ee474f0-0a9f-4c80-bacb-6140ac9c571c/volumes" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.620522 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.622717 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.627394 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n5fwv" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.627902 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.628968 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.634607 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.681169 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.683283 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.683956 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f1d9efee-4efc-460c-ba99-36df95d96a48","Type":"ContainerStarted","Data":"05907e4cf219f001dc32e5a20297c2b77fdfa1103255b11580a6ce5f55551597"} Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.690385 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.694186 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.695662 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerStarted","Data":"86e1161727ab2cd73b072fdf519043d6472e246c487e974f9831da74a28761ca"} Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.698928 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-kvl85" event={"ID":"73149717-cef1-4092-b8af-5cceb0cd4830","Type":"ContainerStarted","Data":"4a44b5d58cdf5feb94a5580fb9e450c1e9bc2b9d16727c60783a8d272f3470a4"} Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.704642 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nj6dc" event={"ID":"a560c2a6-316e-4d1c-8b84-4a44aec7e772","Type":"ContainerStarted","Data":"0ce4edff93f5c6b558eda4f8d79c20719809371092b22ec4240d53027338cab9"} Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.716495 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t9c5x"] Nov 22 08:24:37 crc kubenswrapper[4735]: E1122 08:24:37.717169 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-qsp4r" podUID="c9f75ef9-02e1-4af0-b532-2e744d02d842" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.737184 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nj6dc" podStartSLOduration=2.869911542 podStartE2EDuration="30.737166141s" podCreationTimestamp="2025-11-22 08:24:07 +0000 UTC" firstStartedPulling="2025-11-22 08:24:09.44232819 +0000 UTC m=+1271.046666795" lastFinishedPulling="2025-11-22 08:24:37.309582789 +0000 UTC m=+1298.913921394" observedRunningTime="2025-11-22 08:24:37.726809883 +0000 UTC m=+1299.331148488" watchObservedRunningTime="2025-11-22 08:24:37.737166141 +0000 UTC m=+1299.341504746" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.752393 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-kvl85" podStartSLOduration=2.483806313 podStartE2EDuration="30.752373214s" podCreationTimestamp="2025-11-22 08:24:07 +0000 UTC" firstStartedPulling="2025-11-22 08:24:08.846620932 +0000 UTC m=+1270.450959537" lastFinishedPulling="2025-11-22 08:24:37.115187833 +0000 UTC m=+1298.719526438" observedRunningTime="2025-11-22 08:24:37.738851288 +0000 UTC m=+1299.343189903" watchObservedRunningTime="2025-11-22 08:24:37.752373214 +0000 UTC m=+1299.356711819" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756091 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756182 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756208 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756313 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756526 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-logs\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756560 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756675 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756745 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756792 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.756817 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tktc\" (UniqueName: \"kubernetes.io/projected/c861c1dd-89a1-42d7-a94e-06d62768fe27-kube-api-access-9tktc\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.759621 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5j4v\" (UniqueName: \"kubernetes.io/projected/14edb18c-068d-446e-a291-eab6189d49e6-kube-api-access-v5j4v\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.759678 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-logs\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.759704 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.759755 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.780859 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=46.780830185 podStartE2EDuration="46.780830185s" podCreationTimestamp="2025-11-22 08:23:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:37.769353886 +0000 UTC m=+1299.373692501" watchObservedRunningTime="2025-11-22 08:24:37.780830185 +0000 UTC m=+1299.385168790" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.821713 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-f4gcs"] Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.862920 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863019 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tktc\" (UniqueName: \"kubernetes.io/projected/c861c1dd-89a1-42d7-a94e-06d62768fe27-kube-api-access-9tktc\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863070 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863146 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5j4v\" (UniqueName: \"kubernetes.io/projected/14edb18c-068d-446e-a291-eab6189d49e6-kube-api-access-v5j4v\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863303 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-logs\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863322 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863345 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863390 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863419 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863435 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863496 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863573 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-logs\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863599 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.863661 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.864161 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-logs\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.864873 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.865378 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.865940 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.866860 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.871158 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.874324 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.881657 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-logs\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.881833 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.882640 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.887136 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.887538 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.888030 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5j4v\" (UniqueName: \"kubernetes.io/projected/14edb18c-068d-446e-a291-eab6189d49e6-kube-api-access-v5j4v\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.906021 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tktc\" (UniqueName: \"kubernetes.io/projected/c861c1dd-89a1-42d7-a94e-06d62768fe27-kube-api-access-9tktc\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.954152 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:37 crc kubenswrapper[4735]: I1122 08:24:37.968605 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.014645 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.254339 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.668582 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:38 crc kubenswrapper[4735]: W1122 08:24:38.678132 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc861c1dd_89a1_42d7_a94e_06d62768fe27.slice/crio-5c3869f35ae124d3b5ee558f03c6d6b8385dcd1d7780e3e1bb265eb28322c1f0 WatchSource:0}: Error finding container 5c3869f35ae124d3b5ee558f03c6d6b8385dcd1d7780e3e1bb265eb28322c1f0: Status 404 returned error can't find the container with id 5c3869f35ae124d3b5ee558f03c6d6b8385dcd1d7780e3e1bb265eb28322c1f0 Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.747807 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9c5x" event={"ID":"a32d2023-853f-40f2-9322-155b32c82fdb","Type":"ContainerStarted","Data":"fa99cdc915a3e62de9d4cb9945e176041e12d5b45dc2555fe50a48563aeccab1"} Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.747854 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9c5x" event={"ID":"a32d2023-853f-40f2-9322-155b32c82fdb","Type":"ContainerStarted","Data":"27f07786cd50fd15ea9b2f5455e2d50eeb57f38e055d3284aac06b4ec3d95bb2"} Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.751784 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c861c1dd-89a1-42d7-a94e-06d62768fe27","Type":"ContainerStarted","Data":"5c3869f35ae124d3b5ee558f03c6d6b8385dcd1d7780e3e1bb265eb28322c1f0"} Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.754217 4735 generic.go:334] "Generic (PLEG): container finished" podID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" containerID="7321ae022b3ccecbef5306ccc3c4f3d5f59f563b6a68a82dca970b37f97ca13d" exitCode=0 Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.755155 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" event={"ID":"bd32522c-f1ea-4f33-9e30-7c574e3b427d","Type":"ContainerDied","Data":"7321ae022b3ccecbef5306ccc3c4f3d5f59f563b6a68a82dca970b37f97ca13d"} Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.755182 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" event={"ID":"bd32522c-f1ea-4f33-9e30-7c574e3b427d","Type":"ContainerStarted","Data":"9fa88659d8625c54db900145c186bde75a86b1f741570d14570bda6dae6d0c31"} Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.785404 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-t9c5x" podStartSLOduration=17.785379924 podStartE2EDuration="17.785379924s" podCreationTimestamp="2025-11-22 08:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:38.768035982 +0000 UTC m=+1300.372374587" watchObservedRunningTime="2025-11-22 08:24:38.785379924 +0000 UTC m=+1300.389718519" Nov 22 08:24:38 crc kubenswrapper[4735]: I1122 08:24:38.929071 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:39 crc kubenswrapper[4735]: I1122 08:24:39.779747 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c861c1dd-89a1-42d7-a94e-06d62768fe27","Type":"ContainerStarted","Data":"134a3fe6dbe39edd37b72f6318fd561fbf642c5e35fbae6c588f3735179d9dcc"} Nov 22 08:24:39 crc kubenswrapper[4735]: I1122 08:24:39.791811 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14edb18c-068d-446e-a291-eab6189d49e6","Type":"ContainerStarted","Data":"065e4bcdac6bd4f3d0292eb3eb4ef1295c9a6d801bb886f052e5e127b5b3a9a4"} Nov 22 08:24:39 crc kubenswrapper[4735]: I1122 08:24:39.800601 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" event={"ID":"bd32522c-f1ea-4f33-9e30-7c574e3b427d","Type":"ContainerStarted","Data":"b06540d3d33454a9a5468d54e1de0ff74acd3059e5e9cddefb6225eec62d7681"} Nov 22 08:24:39 crc kubenswrapper[4735]: I1122 08:24:39.800851 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:39 crc kubenswrapper[4735]: I1122 08:24:39.833968 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" podStartSLOduration=3.833951257 podStartE2EDuration="3.833951257s" podCreationTimestamp="2025-11-22 08:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:39.82220445 +0000 UTC m=+1301.426543055" watchObservedRunningTime="2025-11-22 08:24:39.833951257 +0000 UTC m=+1301.438289862" Nov 22 08:24:39 crc kubenswrapper[4735]: I1122 08:24:39.854363 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:39 crc kubenswrapper[4735]: I1122 08:24:39.926258 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:41 crc kubenswrapper[4735]: I1122 08:24:41.823330 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14edb18c-068d-446e-a291-eab6189d49e6","Type":"ContainerStarted","Data":"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7"} Nov 22 08:24:41 crc kubenswrapper[4735]: I1122 08:24:41.827439 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerStarted","Data":"e8b5d8776713807502178db7368884a6b93fe4c2c627a0b5fedec2752f76bab6"} Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.204243 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.841097 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c861c1dd-89a1-42d7-a94e-06d62768fe27","Type":"ContainerStarted","Data":"8f626e28b82b7a010070b3c46d6987197699f306a73d9f8768a256b969462313"} Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.841211 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerName="glance-log" containerID="cri-o://134a3fe6dbe39edd37b72f6318fd561fbf642c5e35fbae6c588f3735179d9dcc" gracePeriod=30 Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.841249 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerName="glance-httpd" containerID="cri-o://8f626e28b82b7a010070b3c46d6987197699f306a73d9f8768a256b969462313" gracePeriod=30 Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.846496 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14edb18c-068d-446e-a291-eab6189d49e6","Type":"ContainerStarted","Data":"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e"} Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.846653 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="14edb18c-068d-446e-a291-eab6189d49e6" containerName="glance-log" containerID="cri-o://48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7" gracePeriod=30 Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.846800 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="14edb18c-068d-446e-a291-eab6189d49e6" containerName="glance-httpd" containerID="cri-o://7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e" gracePeriod=30 Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.929423 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.929271754 podStartE2EDuration="6.929271754s" podCreationTimestamp="2025-11-22 08:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:42.863051492 +0000 UTC m=+1304.467390097" watchObservedRunningTime="2025-11-22 08:24:42.929271754 +0000 UTC m=+1304.533610359" Nov 22 08:24:42 crc kubenswrapper[4735]: I1122 08:24:42.932697 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.932678359 podStartE2EDuration="6.932678359s" podCreationTimestamp="2025-11-22 08:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:42.904370502 +0000 UTC m=+1304.508709107" watchObservedRunningTime="2025-11-22 08:24:42.932678359 +0000 UTC m=+1304.537016964" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.651708 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.765661 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5j4v\" (UniqueName: \"kubernetes.io/projected/14edb18c-068d-446e-a291-eab6189d49e6-kube-api-access-v5j4v\") pod \"14edb18c-068d-446e-a291-eab6189d49e6\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.765733 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-httpd-run\") pod \"14edb18c-068d-446e-a291-eab6189d49e6\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.765789 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-config-data\") pod \"14edb18c-068d-446e-a291-eab6189d49e6\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.765807 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-scripts\") pod \"14edb18c-068d-446e-a291-eab6189d49e6\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.765855 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-logs\") pod \"14edb18c-068d-446e-a291-eab6189d49e6\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.765933 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"14edb18c-068d-446e-a291-eab6189d49e6\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.765994 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-combined-ca-bundle\") pod \"14edb18c-068d-446e-a291-eab6189d49e6\" (UID: \"14edb18c-068d-446e-a291-eab6189d49e6\") " Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.767029 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "14edb18c-068d-446e-a291-eab6189d49e6" (UID: "14edb18c-068d-446e-a291-eab6189d49e6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.767439 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-logs" (OuterVolumeSpecName: "logs") pod "14edb18c-068d-446e-a291-eab6189d49e6" (UID: "14edb18c-068d-446e-a291-eab6189d49e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.772745 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "14edb18c-068d-446e-a291-eab6189d49e6" (UID: "14edb18c-068d-446e-a291-eab6189d49e6"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.772748 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14edb18c-068d-446e-a291-eab6189d49e6-kube-api-access-v5j4v" (OuterVolumeSpecName: "kube-api-access-v5j4v") pod "14edb18c-068d-446e-a291-eab6189d49e6" (UID: "14edb18c-068d-446e-a291-eab6189d49e6"). InnerVolumeSpecName "kube-api-access-v5j4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.774565 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-scripts" (OuterVolumeSpecName: "scripts") pod "14edb18c-068d-446e-a291-eab6189d49e6" (UID: "14edb18c-068d-446e-a291-eab6189d49e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.801590 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14edb18c-068d-446e-a291-eab6189d49e6" (UID: "14edb18c-068d-446e-a291-eab6189d49e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.835287 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-config-data" (OuterVolumeSpecName: "config-data") pod "14edb18c-068d-446e-a291-eab6189d49e6" (UID: "14edb18c-068d-446e-a291-eab6189d49e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.861625 4735 generic.go:334] "Generic (PLEG): container finished" podID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerID="8f626e28b82b7a010070b3c46d6987197699f306a73d9f8768a256b969462313" exitCode=0 Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.861668 4735 generic.go:334] "Generic (PLEG): container finished" podID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerID="134a3fe6dbe39edd37b72f6318fd561fbf642c5e35fbae6c588f3735179d9dcc" exitCode=143 Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.861709 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c861c1dd-89a1-42d7-a94e-06d62768fe27","Type":"ContainerDied","Data":"8f626e28b82b7a010070b3c46d6987197699f306a73d9f8768a256b969462313"} Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.861751 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c861c1dd-89a1-42d7-a94e-06d62768fe27","Type":"ContainerDied","Data":"134a3fe6dbe39edd37b72f6318fd561fbf642c5e35fbae6c588f3735179d9dcc"} Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.864747 4735 generic.go:334] "Generic (PLEG): container finished" podID="14edb18c-068d-446e-a291-eab6189d49e6" containerID="7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e" exitCode=0 Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.864782 4735 generic.go:334] "Generic (PLEG): container finished" podID="14edb18c-068d-446e-a291-eab6189d49e6" containerID="48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7" exitCode=143 Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.864833 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14edb18c-068d-446e-a291-eab6189d49e6","Type":"ContainerDied","Data":"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e"} Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.864859 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.864877 4735 scope.go:117] "RemoveContainer" containerID="7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.864866 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14edb18c-068d-446e-a291-eab6189d49e6","Type":"ContainerDied","Data":"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7"} Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.865009 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14edb18c-068d-446e-a291-eab6189d49e6","Type":"ContainerDied","Data":"065e4bcdac6bd4f3d0292eb3eb4ef1295c9a6d801bb886f052e5e127b5b3a9a4"} Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871049 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871073 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871084 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871121 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871131 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14edb18c-068d-446e-a291-eab6189d49e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871148 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5j4v\" (UniqueName: \"kubernetes.io/projected/14edb18c-068d-446e-a291-eab6189d49e6-kube-api-access-v5j4v\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871160 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14edb18c-068d-446e-a291-eab6189d49e6-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871799 4735 generic.go:334] "Generic (PLEG): container finished" podID="a560c2a6-316e-4d1c-8b84-4a44aec7e772" containerID="0ce4edff93f5c6b558eda4f8d79c20719809371092b22ec4240d53027338cab9" exitCode=0 Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.871828 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nj6dc" event={"ID":"a560c2a6-316e-4d1c-8b84-4a44aec7e772","Type":"ContainerDied","Data":"0ce4edff93f5c6b558eda4f8d79c20719809371092b22ec4240d53027338cab9"} Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.927574 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.934691 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.940172 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.971978 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:43 crc kubenswrapper[4735]: E1122 08:24:43.972425 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14edb18c-068d-446e-a291-eab6189d49e6" containerName="glance-log" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.972440 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="14edb18c-068d-446e-a291-eab6189d49e6" containerName="glance-log" Nov 22 08:24:43 crc kubenswrapper[4735]: E1122 08:24:43.972485 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14edb18c-068d-446e-a291-eab6189d49e6" containerName="glance-httpd" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.972492 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="14edb18c-068d-446e-a291-eab6189d49e6" containerName="glance-httpd" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.972695 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="14edb18c-068d-446e-a291-eab6189d49e6" containerName="glance-log" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.972716 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="14edb18c-068d-446e-a291-eab6189d49e6" containerName="glance-httpd" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.973793 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.974329 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.976842 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.981034 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 22 08:24:43 crc kubenswrapper[4735]: I1122 08:24:43.981493 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.076754 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.076814 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-logs\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.076846 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-scripts\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.076886 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.076924 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl9kn\" (UniqueName: \"kubernetes.io/projected/701597f5-e695-485b-a08c-cec90dad760f-kube-api-access-xl9kn\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.076951 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-config-data\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.076984 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.077011 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.178692 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.178747 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl9kn\" (UniqueName: \"kubernetes.io/projected/701597f5-e695-485b-a08c-cec90dad760f-kube-api-access-xl9kn\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.178767 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-config-data\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.178791 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.178815 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.178975 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.179001 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-logs\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.179021 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-scripts\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.179870 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.181340 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.181818 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-logs\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.185304 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-scripts\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.186940 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.189763 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-config-data\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.206670 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.206733 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl9kn\" (UniqueName: \"kubernetes.io/projected/701597f5-e695-485b-a08c-cec90dad760f-kube-api-access-xl9kn\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.221783 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.303667 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.883360 4735 generic.go:334] "Generic (PLEG): container finished" podID="a32d2023-853f-40f2-9322-155b32c82fdb" containerID="fa99cdc915a3e62de9d4cb9945e176041e12d5b45dc2555fe50a48563aeccab1" exitCode=0 Nov 22 08:24:44 crc kubenswrapper[4735]: I1122 08:24:44.883425 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9c5x" event={"ID":"a32d2023-853f-40f2-9322-155b32c82fdb","Type":"ContainerDied","Data":"fa99cdc915a3e62de9d4cb9945e176041e12d5b45dc2555fe50a48563aeccab1"} Nov 22 08:24:45 crc kubenswrapper[4735]: I1122 08:24:45.296098 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14edb18c-068d-446e-a291-eab6189d49e6" path="/var/lib/kubelet/pods/14edb18c-068d-446e-a291-eab6189d49e6/volumes" Nov 22 08:24:45 crc kubenswrapper[4735]: I1122 08:24:45.924708 4735 generic.go:334] "Generic (PLEG): container finished" podID="73149717-cef1-4092-b8af-5cceb0cd4830" containerID="4a44b5d58cdf5feb94a5580fb9e450c1e9bc2b9d16727c60783a8d272f3470a4" exitCode=0 Nov 22 08:24:45 crc kubenswrapper[4735]: I1122 08:24:45.924796 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-kvl85" event={"ID":"73149717-cef1-4092-b8af-5cceb0cd4830","Type":"ContainerDied","Data":"4a44b5d58cdf5feb94a5580fb9e450c1e9bc2b9d16727c60783a8d272f3470a4"} Nov 22 08:24:45 crc kubenswrapper[4735]: I1122 08:24:45.928069 4735 generic.go:334] "Generic (PLEG): container finished" podID="f525dad0-8aeb-4928-b87f-561f2c3f3228" containerID="427d0f580305583837f5847a430716a462eec9578979f975200112fc0c9a452c" exitCode=0 Nov 22 08:24:45 crc kubenswrapper[4735]: I1122 08:24:45.928260 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qptkr" event={"ID":"f525dad0-8aeb-4928-b87f-561f2c3f3228","Type":"ContainerDied","Data":"427d0f580305583837f5847a430716a462eec9578979f975200112fc0c9a452c"} Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.131186 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.131645 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.131716 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.132523 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4afcbebd71a886d2b368e2d15dd469e8f84d1fd0d397678cdfa5ebd38afdfac1"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.132596 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://4afcbebd71a886d2b368e2d15dd469e8f84d1fd0d397678cdfa5ebd38afdfac1" gracePeriod=600 Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.709187 4735 scope.go:117] "RemoveContainer" containerID="48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7" Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.934677 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.964659 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nj6dc" event={"ID":"a560c2a6-316e-4d1c-8b84-4a44aec7e772","Type":"ContainerDied","Data":"e9bc50eb9eb0e54d229d4e7b1f9e9d1840500a3aeedfd5da01397ba1879112d5"} Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.964711 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9bc50eb9eb0e54d229d4e7b1f9e9d1840500a3aeedfd5da01397ba1879112d5" Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.967931 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9c5x" event={"ID":"a32d2023-853f-40f2-9322-155b32c82fdb","Type":"ContainerDied","Data":"27f07786cd50fd15ea9b2f5455e2d50eeb57f38e055d3284aac06b4ec3d95bb2"} Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.967997 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27f07786cd50fd15ea9b2f5455e2d50eeb57f38e055d3284aac06b4ec3d95bb2" Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.976911 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="4afcbebd71a886d2b368e2d15dd469e8f84d1fd0d397678cdfa5ebd38afdfac1" exitCode=0 Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.976972 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"4afcbebd71a886d2b368e2d15dd469e8f84d1fd0d397678cdfa5ebd38afdfac1"} Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.982860 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c861c1dd-89a1-42d7-a94e-06d62768fe27","Type":"ContainerDied","Data":"5c3869f35ae124d3b5ee558f03c6d6b8385dcd1d7780e3e1bb265eb28322c1f0"} Nov 22 08:24:46 crc kubenswrapper[4735]: I1122 08:24:46.982911 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c3869f35ae124d3b5ee558f03c6d6b8385dcd1d7780e3e1bb265eb28322c1f0" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.007201 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-7cmlv"] Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.012197 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" podUID="e6c6422b-0bb3-4c77-82dc-540f707ab181" containerName="dnsmasq-dns" containerID="cri-o://ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80" gracePeriod=10 Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.048113 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.058235 4735 scope.go:117] "RemoveContainer" containerID="7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e" Nov 22 08:24:47 crc kubenswrapper[4735]: E1122 08:24:47.059720 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e\": container with ID starting with 7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e not found: ID does not exist" containerID="7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.059751 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e"} err="failed to get container status \"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e\": rpc error: code = NotFound desc = could not find container \"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e\": container with ID starting with 7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e not found: ID does not exist" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.059772 4735 scope.go:117] "RemoveContainer" containerID="48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7" Nov 22 08:24:47 crc kubenswrapper[4735]: E1122 08:24:47.060660 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7\": container with ID starting with 48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7 not found: ID does not exist" containerID="48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.060678 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7"} err="failed to get container status \"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7\": rpc error: code = NotFound desc = could not find container \"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7\": container with ID starting with 48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7 not found: ID does not exist" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.060694 4735 scope.go:117] "RemoveContainer" containerID="7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.061509 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e"} err="failed to get container status \"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e\": rpc error: code = NotFound desc = could not find container \"7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e\": container with ID starting with 7dbff4c2dc75bcee6a5202aa149398bc093fc4af938ebca7697ba0a5198f8f4e not found: ID does not exist" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.061537 4735 scope.go:117] "RemoveContainer" containerID="48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.061987 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7"} err="failed to get container status \"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7\": rpc error: code = NotFound desc = could not find container \"48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7\": container with ID starting with 48e6bcb0a93faca3f01fa43dab921935e144765d46844020ced5be05cf852eb7 not found: ID does not exist" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.062038 4735 scope.go:117] "RemoveContainer" containerID="33c0e7ebc11dc73bdfa42a0d62c634dcd145b69c651d39f5221e590919439f72" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.128060 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.146792 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-scripts\") pod \"a32d2023-853f-40f2-9322-155b32c82fdb\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.146921 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghxxd\" (UniqueName: \"kubernetes.io/projected/a32d2023-853f-40f2-9322-155b32c82fdb-kube-api-access-ghxxd\") pod \"a32d2023-853f-40f2-9322-155b32c82fdb\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.146978 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"c861c1dd-89a1-42d7-a94e-06d62768fe27\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147028 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-combined-ca-bundle\") pod \"c861c1dd-89a1-42d7-a94e-06d62768fe27\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147088 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-config-data\") pod \"a32d2023-853f-40f2-9322-155b32c82fdb\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147116 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tktc\" (UniqueName: \"kubernetes.io/projected/c861c1dd-89a1-42d7-a94e-06d62768fe27-kube-api-access-9tktc\") pod \"c861c1dd-89a1-42d7-a94e-06d62768fe27\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147139 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-combined-ca-bundle\") pod \"a32d2023-853f-40f2-9322-155b32c82fdb\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147172 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-scripts\") pod \"c861c1dd-89a1-42d7-a94e-06d62768fe27\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147194 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-logs\") pod \"c861c1dd-89a1-42d7-a94e-06d62768fe27\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147239 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-fernet-keys\") pod \"a32d2023-853f-40f2-9322-155b32c82fdb\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147263 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-config-data\") pod \"c861c1dd-89a1-42d7-a94e-06d62768fe27\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147296 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-httpd-run\") pod \"c861c1dd-89a1-42d7-a94e-06d62768fe27\" (UID: \"c861c1dd-89a1-42d7-a94e-06d62768fe27\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.147374 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-credential-keys\") pod \"a32d2023-853f-40f2-9322-155b32c82fdb\" (UID: \"a32d2023-853f-40f2-9322-155b32c82fdb\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.148077 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-logs" (OuterVolumeSpecName: "logs") pod "c861c1dd-89a1-42d7-a94e-06d62768fe27" (UID: "c861c1dd-89a1-42d7-a94e-06d62768fe27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.148311 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.148645 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c861c1dd-89a1-42d7-a94e-06d62768fe27" (UID: "c861c1dd-89a1-42d7-a94e-06d62768fe27"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.155436 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a32d2023-853f-40f2-9322-155b32c82fdb" (UID: "a32d2023-853f-40f2-9322-155b32c82fdb"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.157324 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a32d2023-853f-40f2-9322-155b32c82fdb" (UID: "a32d2023-853f-40f2-9322-155b32c82fdb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.157539 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c861c1dd-89a1-42d7-a94e-06d62768fe27-kube-api-access-9tktc" (OuterVolumeSpecName: "kube-api-access-9tktc") pod "c861c1dd-89a1-42d7-a94e-06d62768fe27" (UID: "c861c1dd-89a1-42d7-a94e-06d62768fe27"). InnerVolumeSpecName "kube-api-access-9tktc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.165218 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-scripts" (OuterVolumeSpecName: "scripts") pod "c861c1dd-89a1-42d7-a94e-06d62768fe27" (UID: "c861c1dd-89a1-42d7-a94e-06d62768fe27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.165260 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-scripts" (OuterVolumeSpecName: "scripts") pod "a32d2023-853f-40f2-9322-155b32c82fdb" (UID: "a32d2023-853f-40f2-9322-155b32c82fdb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.166357 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "c861c1dd-89a1-42d7-a94e-06d62768fe27" (UID: "c861c1dd-89a1-42d7-a94e-06d62768fe27"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.167412 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32d2023-853f-40f2-9322-155b32c82fdb-kube-api-access-ghxxd" (OuterVolumeSpecName: "kube-api-access-ghxxd") pod "a32d2023-853f-40f2-9322-155b32c82fdb" (UID: "a32d2023-853f-40f2-9322-155b32c82fdb"). InnerVolumeSpecName "kube-api-access-ghxxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.175672 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.249775 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-config-data\") pod \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.249851 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a560c2a6-316e-4d1c-8b84-4a44aec7e772-logs\") pod \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.249916 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqbrl\" (UniqueName: \"kubernetes.io/projected/a560c2a6-316e-4d1c-8b84-4a44aec7e772-kube-api-access-tqbrl\") pod \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250066 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-combined-ca-bundle\") pod \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250167 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-scripts\") pod \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\" (UID: \"a560c2a6-316e-4d1c-8b84-4a44aec7e772\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250614 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250634 4735 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250647 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c861c1dd-89a1-42d7-a94e-06d62768fe27-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250655 4735 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250664 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250672 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghxxd\" (UniqueName: \"kubernetes.io/projected/a32d2023-853f-40f2-9322-155b32c82fdb-kube-api-access-ghxxd\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250694 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250703 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tktc\" (UniqueName: \"kubernetes.io/projected/c861c1dd-89a1-42d7-a94e-06d62768fe27-kube-api-access-9tktc\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.250907 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a560c2a6-316e-4d1c-8b84-4a44aec7e772-logs" (OuterVolumeSpecName: "logs") pod "a560c2a6-316e-4d1c-8b84-4a44aec7e772" (UID: "a560c2a6-316e-4d1c-8b84-4a44aec7e772"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.254482 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a560c2a6-316e-4d1c-8b84-4a44aec7e772-kube-api-access-tqbrl" (OuterVolumeSpecName: "kube-api-access-tqbrl") pod "a560c2a6-316e-4d1c-8b84-4a44aec7e772" (UID: "a560c2a6-316e-4d1c-8b84-4a44aec7e772"). InnerVolumeSpecName "kube-api-access-tqbrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.259391 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-scripts" (OuterVolumeSpecName: "scripts") pod "a560c2a6-316e-4d1c-8b84-4a44aec7e772" (UID: "a560c2a6-316e-4d1c-8b84-4a44aec7e772"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.268663 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-config-data" (OuterVolumeSpecName: "config-data") pod "a32d2023-853f-40f2-9322-155b32c82fdb" (UID: "a32d2023-853f-40f2-9322-155b32c82fdb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.286693 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c861c1dd-89a1-42d7-a94e-06d62768fe27" (UID: "c861c1dd-89a1-42d7-a94e-06d62768fe27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.287194 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a32d2023-853f-40f2-9322-155b32c82fdb" (UID: "a32d2023-853f-40f2-9322-155b32c82fdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.299605 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-config-data" (OuterVolumeSpecName: "config-data") pod "a560c2a6-316e-4d1c-8b84-4a44aec7e772" (UID: "a560c2a6-316e-4d1c-8b84-4a44aec7e772"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.309480 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.338556 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a560c2a6-316e-4d1c-8b84-4a44aec7e772" (UID: "a560c2a6-316e-4d1c-8b84-4a44aec7e772"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351827 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351860 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351873 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351884 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351897 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a560c2a6-316e-4d1c-8b84-4a44aec7e772-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351907 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a560c2a6-316e-4d1c-8b84-4a44aec7e772-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351918 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351928 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32d2023-853f-40f2-9322-155b32c82fdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.351939 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqbrl\" (UniqueName: \"kubernetes.io/projected/a560c2a6-316e-4d1c-8b84-4a44aec7e772-kube-api-access-tqbrl\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.382962 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-config-data" (OuterVolumeSpecName: "config-data") pod "c861c1dd-89a1-42d7-a94e-06d62768fe27" (UID: "c861c1dd-89a1-42d7-a94e-06d62768fe27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.425948 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.453719 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nblfx\" (UniqueName: \"kubernetes.io/projected/73149717-cef1-4092-b8af-5cceb0cd4830-kube-api-access-nblfx\") pod \"73149717-cef1-4092-b8af-5cceb0cd4830\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.453878 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-config-data\") pod \"73149717-cef1-4092-b8af-5cceb0cd4830\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.454066 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-combined-ca-bundle\") pod \"73149717-cef1-4092-b8af-5cceb0cd4830\" (UID: \"73149717-cef1-4092-b8af-5cceb0cd4830\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.454569 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c861c1dd-89a1-42d7-a94e-06d62768fe27-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.457740 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73149717-cef1-4092-b8af-5cceb0cd4830-kube-api-access-nblfx" (OuterVolumeSpecName: "kube-api-access-nblfx") pod "73149717-cef1-4092-b8af-5cceb0cd4830" (UID: "73149717-cef1-4092-b8af-5cceb0cd4830"). InnerVolumeSpecName "kube-api-access-nblfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.495066 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73149717-cef1-4092-b8af-5cceb0cd4830" (UID: "73149717-cef1-4092-b8af-5cceb0cd4830"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.558813 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.558858 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nblfx\" (UniqueName: \"kubernetes.io/projected/73149717-cef1-4092-b8af-5cceb0cd4830-kube-api-access-nblfx\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.583812 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-config-data" (OuterVolumeSpecName: "config-data") pod "73149717-cef1-4092-b8af-5cceb0cd4830" (UID: "73149717-cef1-4092-b8af-5cceb0cd4830"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.596032 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.638971 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.659840 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7tt4\" (UniqueName: \"kubernetes.io/projected/f525dad0-8aeb-4928-b87f-561f2c3f3228-kube-api-access-m7tt4\") pod \"f525dad0-8aeb-4928-b87f-561f2c3f3228\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.659879 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-combined-ca-bundle\") pod \"f525dad0-8aeb-4928-b87f-561f2c3f3228\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.659952 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tx42\" (UniqueName: \"kubernetes.io/projected/e6c6422b-0bb3-4c77-82dc-540f707ab181-kube-api-access-2tx42\") pod \"e6c6422b-0bb3-4c77-82dc-540f707ab181\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.660027 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-sb\") pod \"e6c6422b-0bb3-4c77-82dc-540f707ab181\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.660063 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-swift-storage-0\") pod \"e6c6422b-0bb3-4c77-82dc-540f707ab181\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.660115 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-nb\") pod \"e6c6422b-0bb3-4c77-82dc-540f707ab181\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.660147 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-config\") pod \"e6c6422b-0bb3-4c77-82dc-540f707ab181\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.660182 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-svc\") pod \"e6c6422b-0bb3-4c77-82dc-540f707ab181\" (UID: \"e6c6422b-0bb3-4c77-82dc-540f707ab181\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.660214 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-config\") pod \"f525dad0-8aeb-4928-b87f-561f2c3f3228\" (UID: \"f525dad0-8aeb-4928-b87f-561f2c3f3228\") " Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.660688 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149717-cef1-4092-b8af-5cceb0cd4830-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.679938 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6c6422b-0bb3-4c77-82dc-540f707ab181-kube-api-access-2tx42" (OuterVolumeSpecName: "kube-api-access-2tx42") pod "e6c6422b-0bb3-4c77-82dc-540f707ab181" (UID: "e6c6422b-0bb3-4c77-82dc-540f707ab181"). InnerVolumeSpecName "kube-api-access-2tx42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.683559 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f525dad0-8aeb-4928-b87f-561f2c3f3228-kube-api-access-m7tt4" (OuterVolumeSpecName: "kube-api-access-m7tt4") pod "f525dad0-8aeb-4928-b87f-561f2c3f3228" (UID: "f525dad0-8aeb-4928-b87f-561f2c3f3228"). InnerVolumeSpecName "kube-api-access-m7tt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.752620 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-config" (OuterVolumeSpecName: "config") pod "f525dad0-8aeb-4928-b87f-561f2c3f3228" (UID: "f525dad0-8aeb-4928-b87f-561f2c3f3228"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.756335 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6c6422b-0bb3-4c77-82dc-540f707ab181" (UID: "e6c6422b-0bb3-4c77-82dc-540f707ab181"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.758571 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e6c6422b-0bb3-4c77-82dc-540f707ab181" (UID: "e6c6422b-0bb3-4c77-82dc-540f707ab181"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.765196 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7tt4\" (UniqueName: \"kubernetes.io/projected/f525dad0-8aeb-4928-b87f-561f2c3f3228-kube-api-access-m7tt4\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.765230 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tx42\" (UniqueName: \"kubernetes.io/projected/e6c6422b-0bb3-4c77-82dc-540f707ab181-kube-api-access-2tx42\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.765244 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.765258 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.765271 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.770567 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-config" (OuterVolumeSpecName: "config") pod "e6c6422b-0bb3-4c77-82dc-540f707ab181" (UID: "e6c6422b-0bb3-4c77-82dc-540f707ab181"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.778634 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f525dad0-8aeb-4928-b87f-561f2c3f3228" (UID: "f525dad0-8aeb-4928-b87f-561f2c3f3228"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.796758 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e6c6422b-0bb3-4c77-82dc-540f707ab181" (UID: "e6c6422b-0bb3-4c77-82dc-540f707ab181"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.796811 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e6c6422b-0bb3-4c77-82dc-540f707ab181" (UID: "e6c6422b-0bb3-4c77-82dc-540f707ab181"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.813811 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:24:47 crc kubenswrapper[4735]: W1122 08:24:47.820970 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod701597f5_e695_485b_a08c_cec90dad760f.slice/crio-751d33b56e4e45daa63f2dcf6e63d1fe205f89093963f23e7a52db7f4febab2d WatchSource:0}: Error finding container 751d33b56e4e45daa63f2dcf6e63d1fe205f89093963f23e7a52db7f4febab2d: Status 404 returned error can't find the container with id 751d33b56e4e45daa63f2dcf6e63d1fe205f89093963f23e7a52db7f4febab2d Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.867347 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.867381 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f525dad0-8aeb-4928-b87f-561f2c3f3228-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.867400 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:47 crc kubenswrapper[4735]: I1122 08:24:47.867411 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6c6422b-0bb3-4c77-82dc-540f707ab181-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.025093 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerStarted","Data":"1a09894c33583101467d2d5dba0247e78fc51e16cbb109ee8485a72316353cea"} Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.029017 4735 generic.go:334] "Generic (PLEG): container finished" podID="e6c6422b-0bb3-4c77-82dc-540f707ab181" containerID="ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80" exitCode=0 Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.029063 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" event={"ID":"e6c6422b-0bb3-4c77-82dc-540f707ab181","Type":"ContainerDied","Data":"ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80"} Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.029080 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" event={"ID":"e6c6422b-0bb3-4c77-82dc-540f707ab181","Type":"ContainerDied","Data":"964d656c9dbff9fd6ee385c89b07045eb83deefe3f6fd96ad7d00b1c1620d0e7"} Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.029095 4735 scope.go:117] "RemoveContainer" containerID="ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.029187 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-7cmlv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.031922 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"701597f5-e695-485b-a08c-cec90dad760f","Type":"ContainerStarted","Data":"751d33b56e4e45daa63f2dcf6e63d1fe205f89093963f23e7a52db7f4febab2d"} Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.045111 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qptkr" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.046153 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qptkr" event={"ID":"f525dad0-8aeb-4928-b87f-561f2c3f3228","Type":"ContainerDied","Data":"0b9ce23a551ece580c6f82eadd857d34497ca3c41f1b99f1f06eee7cfbad15eb"} Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.046302 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b9ce23a551ece580c6f82eadd857d34497ca3c41f1b99f1f06eee7cfbad15eb" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.076738 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-kvl85" event={"ID":"73149717-cef1-4092-b8af-5cceb0cd4830","Type":"ContainerDied","Data":"d15cfb6d7b46b8afe34c869772320d9c7b3670a66523f1f5da90ea2436f17c37"} Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.076769 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15cfb6d7b46b8afe34c869772320d9c7b3670a66523f1f5da90ea2436f17c37" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.076748 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-kvl85" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.078908 4735 scope.go:117] "RemoveContainer" containerID="90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.086224 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nj6dc" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.086546 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7"} Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.087579 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9c5x" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.088374 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.114626 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-7cmlv"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.136353 4735 scope.go:117] "RemoveContainer" containerID="ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.138645 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80\": container with ID starting with ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80 not found: ID does not exist" containerID="ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.138693 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80"} err="failed to get container status \"ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80\": rpc error: code = NotFound desc = could not find container \"ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80\": container with ID starting with ec6d3225db6270694e189b3c1d0067af6e18d34e0d7282cb7f5bb71421eb6c80 not found: ID does not exist" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.138752 4735 scope.go:117] "RemoveContainer" containerID="90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.141050 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b\": container with ID starting with 90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b not found: ID does not exist" containerID="90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.141081 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b"} err="failed to get container status \"90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b\": rpc error: code = NotFound desc = could not find container \"90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b\": container with ID starting with 90f2ea10c365084b8271eb40a979ab34ed125b4976db81d85ef09ab85739ab4b not found: ID does not exist" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.225815 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-7cmlv"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.393789 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hm5w8"] Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.394254 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a560c2a6-316e-4d1c-8b84-4a44aec7e772" containerName="placement-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394267 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a560c2a6-316e-4d1c-8b84-4a44aec7e772" containerName="placement-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.394278 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerName="glance-httpd" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394284 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerName="glance-httpd" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.394301 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c6422b-0bb3-4c77-82dc-540f707ab181" containerName="dnsmasq-dns" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394309 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c6422b-0bb3-4c77-82dc-540f707ab181" containerName="dnsmasq-dns" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.394325 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32d2023-853f-40f2-9322-155b32c82fdb" containerName="keystone-bootstrap" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394330 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32d2023-853f-40f2-9322-155b32c82fdb" containerName="keystone-bootstrap" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.394343 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerName="glance-log" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394348 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerName="glance-log" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.394355 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f525dad0-8aeb-4928-b87f-561f2c3f3228" containerName="neutron-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394360 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f525dad0-8aeb-4928-b87f-561f2c3f3228" containerName="neutron-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.394373 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73149717-cef1-4092-b8af-5cceb0cd4830" containerName="heat-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394379 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="73149717-cef1-4092-b8af-5cceb0cd4830" containerName="heat-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: E1122 08:24:48.394387 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c6422b-0bb3-4c77-82dc-540f707ab181" containerName="init" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394393 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c6422b-0bb3-4c77-82dc-540f707ab181" containerName="init" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394620 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32d2023-853f-40f2-9322-155b32c82fdb" containerName="keystone-bootstrap" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394634 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f525dad0-8aeb-4928-b87f-561f2c3f3228" containerName="neutron-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394646 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerName="glance-log" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394655 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a560c2a6-316e-4d1c-8b84-4a44aec7e772" containerName="placement-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394665 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="73149717-cef1-4092-b8af-5cceb0cd4830" containerName="heat-db-sync" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394677 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" containerName="glance-httpd" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.394685 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6c6422b-0bb3-4c77-82dc-540f707ab181" containerName="dnsmasq-dns" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.395801 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.419127 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5584fd997b-7ztjv"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.423517 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.429864 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hm5w8"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.438013 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.438246 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s4xtc" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.438396 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.438680 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.438927 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.439052 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.449988 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.470543 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5584fd997b-7ztjv"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.494653 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.500705 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-internal-tls-certs\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.501297 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.501334 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-public-tls-certs\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.501596 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.513583 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-combined-ca-bundle\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.513663 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.513830 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-scripts\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.513873 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-config-data\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.513916 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.513949 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-fernet-keys\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.513993 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t9zk\" (UniqueName: \"kubernetes.io/projected/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-kube-api-access-8t9zk\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.514044 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6pf7\" (UniqueName: \"kubernetes.io/projected/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-kube-api-access-b6pf7\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.514125 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-config\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.514230 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-credential-keys\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.523335 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.525911 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.530164 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.533265 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.541538 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.571921 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6fcc565756-vf564"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.573972 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.578896 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fcc565756-vf564"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.582183 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.582492 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.582601 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-c9rr6" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.582705 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.582801 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.627966 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-public-tls-certs\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628013 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628035 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-public-tls-certs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628086 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628142 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-config-data\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628179 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628196 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-combined-ca-bundle\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628216 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628236 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9tdb\" (UniqueName: \"kubernetes.io/projected/14444255-5953-4eef-8a78-46697c99b0e6-kube-api-access-g9tdb\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628279 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-logs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628302 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-scripts\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628316 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hrj7\" (UniqueName: \"kubernetes.io/projected/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-kube-api-access-6hrj7\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628340 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-scripts\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628362 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-config-data\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628385 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628404 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-fernet-keys\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628428 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t9zk\" (UniqueName: \"kubernetes.io/projected/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-kube-api-access-8t9zk\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628447 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628479 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628500 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-internal-tls-certs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628528 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6pf7\" (UniqueName: \"kubernetes.io/projected/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-kube-api-access-b6pf7\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628554 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628583 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-combined-ca-bundle\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628609 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-config\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628656 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-credential-keys\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628699 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628718 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628745 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-internal-tls-certs\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.628766 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.629578 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.630348 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-config\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.630535 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.631620 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.632895 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.666298 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-public-tls-certs\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.670390 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t9zk\" (UniqueName: \"kubernetes.io/projected/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-kube-api-access-8t9zk\") pod \"dnsmasq-dns-55f844cf75-hm5w8\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.680041 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6pf7\" (UniqueName: \"kubernetes.io/projected/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-kube-api-access-b6pf7\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.694238 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-config-data\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.697928 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-fernet-keys\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.698264 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-scripts\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.702710 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-combined-ca-bundle\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.706003 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-credential-keys\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.717877 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4906aa84-20dc-4096-8d7c-0ed5dd4d9103-internal-tls-certs\") pod \"keystone-5584fd997b-7ztjv\" (UID: \"4906aa84-20dc-4096-8d7c-0ed5dd4d9103\") " pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.743978 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744025 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744107 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744123 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-public-tls-certs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744205 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744312 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-config-data\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744389 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9tdb\" (UniqueName: \"kubernetes.io/projected/14444255-5953-4eef-8a78-46697c99b0e6-kube-api-access-g9tdb\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744471 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-logs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744506 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-scripts\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744522 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hrj7\" (UniqueName: \"kubernetes.io/projected/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-kube-api-access-6hrj7\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744620 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744638 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744662 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-internal-tls-certs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744686 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.744719 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-combined-ca-bundle\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.745907 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.746415 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-logs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.748212 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-scripts\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.748959 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.751573 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-796c85cffb-bqt8t"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.756237 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.762007 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.767365 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8sbsw" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.767564 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.768090 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.768594 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.769257 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.774022 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hrj7\" (UniqueName: \"kubernetes.io/projected/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-kube-api-access-6hrj7\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.775531 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-796c85cffb-bqt8t"] Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.780175 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.780582 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.781104 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-public-tls-certs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.781506 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.782170 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-combined-ca-bundle\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.786309 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9tdb\" (UniqueName: \"kubernetes.io/projected/14444255-5953-4eef-8a78-46697c99b0e6-kube-api-access-g9tdb\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.786564 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-config-data\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.786810 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.810141 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbc92e-2676-41c5-abe1-9fc69aa0dad3-internal-tls-certs\") pod \"placement-6fcc565756-vf564\" (UID: \"99fbc92e-2676-41c5-abe1-9fc69aa0dad3\") " pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.810755 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.822921 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.920334 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.947628 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8r6d\" (UniqueName: \"kubernetes.io/projected/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-kube-api-access-s8r6d\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.948020 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-httpd-config\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.948043 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-config\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.948080 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-combined-ca-bundle\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.948150 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-ovndb-tls-certs\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:48 crc kubenswrapper[4735]: I1122 08:24:48.994410 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.053183 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8r6d\" (UniqueName: \"kubernetes.io/projected/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-kube-api-access-s8r6d\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.053423 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-httpd-config\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.053443 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-config\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.053524 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-combined-ca-bundle\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.053696 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-ovndb-tls-certs\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.062876 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-httpd-config\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.063281 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-ovndb-tls-certs\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.068780 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-config\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.072073 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-combined-ca-bundle\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.077194 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8r6d\" (UniqueName: \"kubernetes.io/projected/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-kube-api-access-s8r6d\") pod \"neutron-796c85cffb-bqt8t\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.146264 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"701597f5-e695-485b-a08c-cec90dad760f","Type":"ContainerStarted","Data":"b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a"} Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.307779 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.335352 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c861c1dd-89a1-42d7-a94e-06d62768fe27" path="/var/lib/kubelet/pods/c861c1dd-89a1-42d7-a94e-06d62768fe27/volumes" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.336150 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6c6422b-0bb3-4c77-82dc-540f707ab181" path="/var/lib/kubelet/pods/e6c6422b-0bb3-4c77-82dc-540f707ab181/volumes" Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.522357 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5584fd997b-7ztjv"] Nov 22 08:24:49 crc kubenswrapper[4735]: I1122 08:24:49.818932 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hm5w8"] Nov 22 08:24:50 crc kubenswrapper[4735]: I1122 08:24:50.121496 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fcc565756-vf564"] Nov 22 08:24:50 crc kubenswrapper[4735]: I1122 08:24:50.137498 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:24:50 crc kubenswrapper[4735]: I1122 08:24:50.182623 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z64cd" event={"ID":"e519a2d0-063e-4097-80f8-f4276208317a","Type":"ContainerStarted","Data":"2762bf4a14beb7350f245c7468efcceed103ef8107b01e8b3517f2926729d4f4"} Nov 22 08:24:50 crc kubenswrapper[4735]: I1122 08:24:50.188349 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" event={"ID":"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded","Type":"ContainerStarted","Data":"4099a24b8eefda848208db66f37de6a6daa322fdc076ac7e5aec0157cd4378bf"} Nov 22 08:24:50 crc kubenswrapper[4735]: I1122 08:24:50.191761 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5584fd997b-7ztjv" event={"ID":"4906aa84-20dc-4096-8d7c-0ed5dd4d9103","Type":"ContainerStarted","Data":"3ced875ed039b1b0b3805f9dc84593de0dd15358d3f8883ee202e4c5a63061af"} Nov 22 08:24:50 crc kubenswrapper[4735]: I1122 08:24:50.198175 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-z64cd" podStartSLOduration=3.553509774 podStartE2EDuration="43.198162068s" podCreationTimestamp="2025-11-22 08:24:07 +0000 UTC" firstStartedPulling="2025-11-22 08:24:09.268981648 +0000 UTC m=+1270.873320253" lastFinishedPulling="2025-11-22 08:24:48.913633942 +0000 UTC m=+1310.517972547" observedRunningTime="2025-11-22 08:24:50.197489369 +0000 UTC m=+1311.801827974" watchObservedRunningTime="2025-11-22 08:24:50.198162068 +0000 UTC m=+1311.802500673" Nov 22 08:24:50 crc kubenswrapper[4735]: I1122 08:24:50.211995 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14444255-5953-4eef-8a78-46697c99b0e6","Type":"ContainerStarted","Data":"99d849f553532428f2f4c54b9c2d8252610eb3ecf967bf805ba7ebf712ca620a"} Nov 22 08:24:50 crc kubenswrapper[4735]: I1122 08:24:50.242307 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-796c85cffb-bqt8t"] Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.286329 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.287701 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5584fd997b-7ztjv" event={"ID":"4906aa84-20dc-4096-8d7c-0ed5dd4d9103","Type":"ContainerStarted","Data":"94c09731732bc8ff572a5ff9b95e28aff9c1afb5e0785b8aeee602a463ec8d6e"} Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.294532 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qsp4r" event={"ID":"c9f75ef9-02e1-4af0-b532-2e744d02d842","Type":"ContainerStarted","Data":"6a111f241484eb5391838b9af58a78e8bf22389fb19a1d2438956734070d506d"} Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.303929 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5584fd997b-7ztjv" podStartSLOduration=3.30390693 podStartE2EDuration="3.30390693s" podCreationTimestamp="2025-11-22 08:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:51.300854345 +0000 UTC m=+1312.905192960" watchObservedRunningTime="2025-11-22 08:24:51.30390693 +0000 UTC m=+1312.908245535" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.310164 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-796c85cffb-bqt8t" event={"ID":"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a","Type":"ContainerStarted","Data":"7360b36476c2e0682a7873761f3f256f17cbf7a10797fd7003c2c5fa1c70362e"} Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.310201 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-796c85cffb-bqt8t" event={"ID":"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a","Type":"ContainerStarted","Data":"3f7ca1bb5c0d781ec8e26d6ec48cf79b89e2c7bc60dbbfbb57d274954c011f9a"} Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.314558 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"701597f5-e695-485b-a08c-cec90dad760f","Type":"ContainerStarted","Data":"ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d"} Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.318504 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fcc565756-vf564" event={"ID":"99fbc92e-2676-41c5-abe1-9fc69aa0dad3","Type":"ContainerStarted","Data":"a31bc8571e96d6cf9ffeebf1cf7bd73f3eb25a9d4c93760f8d6b6036d13082c6"} Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.318552 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fcc565756-vf564" event={"ID":"99fbc92e-2676-41c5-abe1-9fc69aa0dad3","Type":"ContainerStarted","Data":"8e7705559a207ef59f8726b83e4b0f781592eeb78cd6040337bd10c479255dca"} Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.323671 4735 generic.go:334] "Generic (PLEG): container finished" podID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" containerID="1ef143c5226e71c9a1ed59a1fe4ca387120a12a958cef498c700805fc6914400" exitCode=0 Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.323726 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" event={"ID":"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded","Type":"ContainerDied","Data":"1ef143c5226e71c9a1ed59a1fe4ca387120a12a958cef498c700805fc6914400"} Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.324402 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-qsp4r" podStartSLOduration=4.307585117 podStartE2EDuration="44.32438057s" podCreationTimestamp="2025-11-22 08:24:07 +0000 UTC" firstStartedPulling="2025-11-22 08:24:08.892056626 +0000 UTC m=+1270.496395231" lastFinishedPulling="2025-11-22 08:24:48.908852079 +0000 UTC m=+1310.513190684" observedRunningTime="2025-11-22 08:24:51.318805595 +0000 UTC m=+1312.923144200" watchObservedRunningTime="2025-11-22 08:24:51.32438057 +0000 UTC m=+1312.928719175" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.364138 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.364115985 podStartE2EDuration="8.364115985s" podCreationTimestamp="2025-11-22 08:24:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:51.347123603 +0000 UTC m=+1312.951462228" watchObservedRunningTime="2025-11-22 08:24:51.364115985 +0000 UTC m=+1312.968454580" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.826935 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7bb54fbc4c-fnr9g"] Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.829539 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.831212 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.831978 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.862136 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bb54fbc4c-fnr9g"] Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.988176 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9tbz\" (UniqueName: \"kubernetes.io/projected/007a6147-937e-485d-9e4c-1a193fdff463-kube-api-access-l9tbz\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.988281 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-ovndb-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.988332 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-httpd-config\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.988417 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-combined-ca-bundle\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.988513 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-public-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.988562 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-internal-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:51 crc kubenswrapper[4735]: I1122 08:24:51.988589 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-config\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.090343 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-combined-ca-bundle\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.090441 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-public-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.090499 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-internal-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.090521 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-config\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.090583 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9tbz\" (UniqueName: \"kubernetes.io/projected/007a6147-937e-485d-9e4c-1a193fdff463-kube-api-access-l9tbz\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.090638 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-ovndb-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.090689 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-httpd-config\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.098314 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-httpd-config\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.098940 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-internal-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.106143 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-config\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.106874 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-public-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.107907 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-combined-ca-bundle\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.108681 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/007a6147-937e-485d-9e4c-1a193fdff463-ovndb-tls-certs\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.131967 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9tbz\" (UniqueName: \"kubernetes.io/projected/007a6147-937e-485d-9e4c-1a193fdff463-kube-api-access-l9tbz\") pod \"neutron-7bb54fbc4c-fnr9g\" (UID: \"007a6147-937e-485d-9e4c-1a193fdff463\") " pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.157576 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.203690 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.210916 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.343596 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14444255-5953-4eef-8a78-46697c99b0e6","Type":"ContainerStarted","Data":"e3cfe12fee68b54c05ce16495c472a50fc6ff4f401cc61a2daa3370f9b6802a5"} Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.343639 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14444255-5953-4eef-8a78-46697c99b0e6","Type":"ContainerStarted","Data":"2bccddfce55cfbff8b4bf1140d2519a32a053a430f02c542e874e26c3b933d3f"} Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.346961 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fcc565756-vf564" event={"ID":"99fbc92e-2676-41c5-abe1-9fc69aa0dad3","Type":"ContainerStarted","Data":"e3e3ece807b81e0417c7b6c862fe57d2928cba7759e498da71df888bb571c36d"} Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.347095 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.349483 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" event={"ID":"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded","Type":"ContainerStarted","Data":"543c4b61014f49d2ae98f0082d0f4a68d4a1fae8f29f3859fcecaaca5f06da64"} Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.350898 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.359649 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-796c85cffb-bqt8t" event={"ID":"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a","Type":"ContainerStarted","Data":"e0cb014851ced8c80e7c5be141d80f1fe266f43982b254effc4846669d5dbbdf"} Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.361049 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.366542 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.406006 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.405981602 podStartE2EDuration="4.405981602s" podCreationTimestamp="2025-11-22 08:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:52.399129401 +0000 UTC m=+1314.003468026" watchObservedRunningTime="2025-11-22 08:24:52.405981602 +0000 UTC m=+1314.010320207" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.459074 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" podStartSLOduration=4.459052398 podStartE2EDuration="4.459052398s" podCreationTimestamp="2025-11-22 08:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:52.455101817 +0000 UTC m=+1314.059440422" watchObservedRunningTime="2025-11-22 08:24:52.459052398 +0000 UTC m=+1314.063391003" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.491403 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-796c85cffb-bqt8t" podStartSLOduration=4.491384926 podStartE2EDuration="4.491384926s" podCreationTimestamp="2025-11-22 08:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:52.475613608 +0000 UTC m=+1314.079952243" watchObservedRunningTime="2025-11-22 08:24:52.491384926 +0000 UTC m=+1314.095723531" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.508469 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6fcc565756-vf564" podStartSLOduration=4.508434001 podStartE2EDuration="4.508434001s" podCreationTimestamp="2025-11-22 08:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:24:52.500424489 +0000 UTC m=+1314.104763094" watchObservedRunningTime="2025-11-22 08:24:52.508434001 +0000 UTC m=+1314.112772606" Nov 22 08:24:52 crc kubenswrapper[4735]: I1122 08:24:52.850393 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bb54fbc4c-fnr9g"] Nov 22 08:24:53 crc kubenswrapper[4735]: I1122 08:24:53.378771 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bb54fbc4c-fnr9g" event={"ID":"007a6147-937e-485d-9e4c-1a193fdff463","Type":"ContainerStarted","Data":"c8d6af9234957a13996d6e17e68d3aae847edb98d42b38d10264082f69ab1ab6"} Nov 22 08:24:53 crc kubenswrapper[4735]: I1122 08:24:53.379151 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fcc565756-vf564" Nov 22 08:24:54 crc kubenswrapper[4735]: I1122 08:24:54.304143 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 22 08:24:54 crc kubenswrapper[4735]: I1122 08:24:54.304207 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 22 08:24:54 crc kubenswrapper[4735]: I1122 08:24:54.361044 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 22 08:24:54 crc kubenswrapper[4735]: I1122 08:24:54.377322 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 22 08:24:54 crc kubenswrapper[4735]: I1122 08:24:54.396741 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bb54fbc4c-fnr9g" event={"ID":"007a6147-937e-485d-9e4c-1a193fdff463","Type":"ContainerStarted","Data":"89d7c7d02ef8dc29d746f89c5f15d479e6e946f6f810927801339751c811d6c6"} Nov 22 08:24:54 crc kubenswrapper[4735]: I1122 08:24:54.398051 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 22 08:24:54 crc kubenswrapper[4735]: I1122 08:24:54.398104 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 22 08:24:56 crc kubenswrapper[4735]: I1122 08:24:56.426893 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:24:58 crc kubenswrapper[4735]: I1122 08:24:58.770708 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:24:58 crc kubenswrapper[4735]: I1122 08:24:58.834014 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-f4gcs"] Nov 22 08:24:58 crc kubenswrapper[4735]: I1122 08:24:58.834273 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" podUID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" containerName="dnsmasq-dns" containerID="cri-o://b06540d3d33454a9a5468d54e1de0ff74acd3059e5e9cddefb6225eec62d7681" gracePeriod=10 Nov 22 08:24:58 crc kubenswrapper[4735]: I1122 08:24:58.921285 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:58 crc kubenswrapper[4735]: I1122 08:24:58.921550 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:58 crc kubenswrapper[4735]: I1122 08:24:58.956779 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:58 crc kubenswrapper[4735]: I1122 08:24:58.979667 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:59 crc kubenswrapper[4735]: I1122 08:24:59.464616 4735 generic.go:334] "Generic (PLEG): container finished" podID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" containerID="b06540d3d33454a9a5468d54e1de0ff74acd3059e5e9cddefb6225eec62d7681" exitCode=0 Nov 22 08:24:59 crc kubenswrapper[4735]: I1122 08:24:59.464683 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" event={"ID":"bd32522c-f1ea-4f33-9e30-7c574e3b427d","Type":"ContainerDied","Data":"b06540d3d33454a9a5468d54e1de0ff74acd3059e5e9cddefb6225eec62d7681"} Nov 22 08:24:59 crc kubenswrapper[4735]: I1122 08:24:59.466266 4735 generic.go:334] "Generic (PLEG): container finished" podID="e519a2d0-063e-4097-80f8-f4276208317a" containerID="2762bf4a14beb7350f245c7468efcceed103ef8107b01e8b3517f2926729d4f4" exitCode=0 Nov 22 08:24:59 crc kubenswrapper[4735]: I1122 08:24:59.467654 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z64cd" event={"ID":"e519a2d0-063e-4097-80f8-f4276208317a","Type":"ContainerDied","Data":"2762bf4a14beb7350f245c7468efcceed103ef8107b01e8b3517f2926729d4f4"} Nov 22 08:24:59 crc kubenswrapper[4735]: I1122 08:24:59.467687 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 22 08:24:59 crc kubenswrapper[4735]: I1122 08:24:59.468186 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 22 08:25:00 crc kubenswrapper[4735]: I1122 08:25:00.943818 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.071885 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-nb\") pod \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.071975 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l4hb\" (UniqueName: \"kubernetes.io/projected/bd32522c-f1ea-4f33-9e30-7c574e3b427d-kube-api-access-4l4hb\") pod \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.072010 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-config\") pod \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.072141 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-swift-storage-0\") pod \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.072160 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-sb\") pod \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.072241 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-svc\") pod \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\" (UID: \"bd32522c-f1ea-4f33-9e30-7c574e3b427d\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.078822 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd32522c-f1ea-4f33-9e30-7c574e3b427d-kube-api-access-4l4hb" (OuterVolumeSpecName: "kube-api-access-4l4hb") pod "bd32522c-f1ea-4f33-9e30-7c574e3b427d" (UID: "bd32522c-f1ea-4f33-9e30-7c574e3b427d"). InnerVolumeSpecName "kube-api-access-4l4hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.136341 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z64cd" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.143893 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.143975 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.147983 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.156252 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bd32522c-f1ea-4f33-9e30-7c574e3b427d" (UID: "bd32522c-f1ea-4f33-9e30-7c574e3b427d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.195075 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l4hb\" (UniqueName: \"kubernetes.io/projected/bd32522c-f1ea-4f33-9e30-7c574e3b427d-kube-api-access-4l4hb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.195127 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.296524 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-db-sync-config-data\") pod \"e519a2d0-063e-4097-80f8-f4276208317a\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.296562 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbxk4\" (UniqueName: \"kubernetes.io/projected/e519a2d0-063e-4097-80f8-f4276208317a-kube-api-access-fbxk4\") pod \"e519a2d0-063e-4097-80f8-f4276208317a\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.296841 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-combined-ca-bundle\") pod \"e519a2d0-063e-4097-80f8-f4276208317a\" (UID: \"e519a2d0-063e-4097-80f8-f4276208317a\") " Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.306919 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-config" (OuterVolumeSpecName: "config") pod "bd32522c-f1ea-4f33-9e30-7c574e3b427d" (UID: "bd32522c-f1ea-4f33-9e30-7c574e3b427d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.311957 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e519a2d0-063e-4097-80f8-f4276208317a-kube-api-access-fbxk4" (OuterVolumeSpecName: "kube-api-access-fbxk4") pod "e519a2d0-063e-4097-80f8-f4276208317a" (UID: "e519a2d0-063e-4097-80f8-f4276208317a"). InnerVolumeSpecName "kube-api-access-fbxk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.312423 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd32522c-f1ea-4f33-9e30-7c574e3b427d" (UID: "bd32522c-f1ea-4f33-9e30-7c574e3b427d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.321128 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e519a2d0-063e-4097-80f8-f4276208317a" (UID: "e519a2d0-063e-4097-80f8-f4276208317a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.353005 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bd32522c-f1ea-4f33-9e30-7c574e3b427d" (UID: "bd32522c-f1ea-4f33-9e30-7c574e3b427d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.379959 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bd32522c-f1ea-4f33-9e30-7c574e3b427d" (UID: "bd32522c-f1ea-4f33-9e30-7c574e3b427d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.400284 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.400316 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.400325 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.400357 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbxk4\" (UniqueName: \"kubernetes.io/projected/e519a2d0-063e-4097-80f8-f4276208317a-kube-api-access-fbxk4\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.400370 4735 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.400383 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd32522c-f1ea-4f33-9e30-7c574e3b427d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.406472 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e519a2d0-063e-4097-80f8-f4276208317a" (UID: "e519a2d0-063e-4097-80f8-f4276208317a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.488127 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z64cd" event={"ID":"e519a2d0-063e-4097-80f8-f4276208317a","Type":"ContainerDied","Data":"9a1ef4f014665f0acfe34cb65044872baa410f461756acfcd9cd9fbc6eb66d60"} Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.488402 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a1ef4f014665f0acfe34cb65044872baa410f461756acfcd9cd9fbc6eb66d60" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.488164 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z64cd" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.490474 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bb54fbc4c-fnr9g" event={"ID":"007a6147-937e-485d-9e4c-1a193fdff463","Type":"ContainerStarted","Data":"a07de1ac79f371aee2ab122ac75bf0fe243e6aade0612b0ba84983c5ff5b2ef8"} Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.490550 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.492938 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" event={"ID":"bd32522c-f1ea-4f33-9e30-7c574e3b427d","Type":"ContainerDied","Data":"9fa88659d8625c54db900145c186bde75a86b1f741570d14570bda6dae6d0c31"} Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.492997 4735 scope.go:117] "RemoveContainer" containerID="b06540d3d33454a9a5468d54e1de0ff74acd3059e5e9cddefb6225eec62d7681" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.493028 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.492968 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-f4gcs" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.493048 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.501737 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e519a2d0-063e-4097-80f8-f4276208317a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.526326 4735 scope.go:117] "RemoveContainer" containerID="7321ae022b3ccecbef5306ccc3c4f3d5f59f563b6a68a82dca970b37f97ca13d" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.532035 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7bb54fbc4c-fnr9g" podStartSLOduration=10.532019336 podStartE2EDuration="10.532019336s" podCreationTimestamp="2025-11-22 08:24:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:01.513154251 +0000 UTC m=+1323.117492856" watchObservedRunningTime="2025-11-22 08:25:01.532019336 +0000 UTC m=+1323.136357941" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.570440 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-f4gcs"] Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.581970 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-f4gcs"] Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.718278 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-66b55c68fd-lfj8p"] Nov 22 08:25:01 crc kubenswrapper[4735]: E1122 08:25:01.719075 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" containerName="dnsmasq-dns" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.719094 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" containerName="dnsmasq-dns" Nov 22 08:25:01 crc kubenswrapper[4735]: E1122 08:25:01.719144 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e519a2d0-063e-4097-80f8-f4276208317a" containerName="barbican-db-sync" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.719151 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e519a2d0-063e-4097-80f8-f4276208317a" containerName="barbican-db-sync" Nov 22 08:25:01 crc kubenswrapper[4735]: E1122 08:25:01.719163 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" containerName="init" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.719170 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" containerName="init" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.719434 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e519a2d0-063e-4097-80f8-f4276208317a" containerName="barbican-db-sync" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.719453 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" containerName="dnsmasq-dns" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.721072 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.725786 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.726120 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.736344 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mk4xm" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.740410 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-66b55c68fd-lfj8p"] Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.798501 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-69b9b656dc-jwxhg"] Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.800346 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.805111 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.813605 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-config-data-custom\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.813779 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-combined-ca-bundle\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.813818 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-config-data\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.813868 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/736e8f87-4b05-4335-954a-47d3c89f84fc-logs\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.813889 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz5cp\" (UniqueName: \"kubernetes.io/projected/736e8f87-4b05-4335-954a-47d3c89f84fc-kube-api-access-rz5cp\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.816845 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69b9b656dc-jwxhg"] Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.845248 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-76v9z"] Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.850942 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.867499 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-76v9z"] Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.917901 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-config-data\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.917962 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-config\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918020 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-config-data-custom\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918058 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918116 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/736e8f87-4b05-4335-954a-47d3c89f84fc-logs\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918135 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz5cp\" (UniqueName: \"kubernetes.io/projected/736e8f87-4b05-4335-954a-47d3c89f84fc-kube-api-access-rz5cp\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918191 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsxs4\" (UniqueName: \"kubernetes.io/projected/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-kube-api-access-rsxs4\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918212 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-config-data-custom\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918294 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918333 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-combined-ca-bundle\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918369 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5ngn\" (UniqueName: \"kubernetes.io/projected/0966ff10-5a13-45a4-a20a-5eb889270c5b-kube-api-access-n5ngn\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918386 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918407 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-combined-ca-bundle\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918436 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0966ff10-5a13-45a4-a20a-5eb889270c5b-logs\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918533 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-config-data\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.918560 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.919079 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/736e8f87-4b05-4335-954a-47d3c89f84fc-logs\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.926126 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-config-data-custom\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.931956 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-combined-ca-bundle\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.934918 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/736e8f87-4b05-4335-954a-47d3c89f84fc-config-data\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.948231 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-85664c95cd-7bnn9"] Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.948668 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz5cp\" (UniqueName: \"kubernetes.io/projected/736e8f87-4b05-4335-954a-47d3c89f84fc-kube-api-access-rz5cp\") pod \"barbican-keystone-listener-66b55c68fd-lfj8p\" (UID: \"736e8f87-4b05-4335-954a-47d3c89f84fc\") " pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.950297 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.954733 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 22 08:25:01 crc kubenswrapper[4735]: I1122 08:25:01.975827 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85664c95cd-7bnn9"] Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020312 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020371 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-combined-ca-bundle\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020406 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data-custom\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020422 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-combined-ca-bundle\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020449 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5ngn\" (UniqueName: \"kubernetes.io/projected/0966ff10-5a13-45a4-a20a-5eb889270c5b-kube-api-access-n5ngn\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020478 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020516 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020531 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0966ff10-5a13-45a4-a20a-5eb889270c5b-logs\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020554 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020580 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-config-data\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020608 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-config\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020628 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020645 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-config-data-custom\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020668 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a80806be-71c0-442d-95df-a8183640736b-logs\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020719 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsxs4\" (UniqueName: \"kubernetes.io/projected/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-kube-api-access-rsxs4\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.020739 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmnfw\" (UniqueName: \"kubernetes.io/projected/a80806be-71c0-442d-95df-a8183640736b-kube-api-access-tmnfw\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.021671 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0966ff10-5a13-45a4-a20a-5eb889270c5b-logs\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.022497 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.022615 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.022879 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.023159 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.023659 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-config\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.026696 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.027982 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-combined-ca-bundle\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.034226 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-config-data\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.038836 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsxs4\" (UniqueName: \"kubernetes.io/projected/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-kube-api-access-rsxs4\") pod \"dnsmasq-dns-85ff748b95-76v9z\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.041033 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0966ff10-5a13-45a4-a20a-5eb889270c5b-config-data-custom\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.042875 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5ngn\" (UniqueName: \"kubernetes.io/projected/0966ff10-5a13-45a4-a20a-5eb889270c5b-kube-api-access-n5ngn\") pod \"barbican-worker-69b9b656dc-jwxhg\" (UID: \"0966ff10-5a13-45a4-a20a-5eb889270c5b\") " pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.063605 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.103139 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.122436 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a80806be-71c0-442d-95df-a8183640736b-logs\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.122531 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmnfw\" (UniqueName: \"kubernetes.io/projected/a80806be-71c0-442d-95df-a8183640736b-kube-api-access-tmnfw\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.122629 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data-custom\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.122648 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-combined-ca-bundle\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.122683 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.123508 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a80806be-71c0-442d-95df-a8183640736b-logs\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.147496 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data-custom\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.148782 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.160478 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmnfw\" (UniqueName: \"kubernetes.io/projected/a80806be-71c0-442d-95df-a8183640736b-kube-api-access-tmnfw\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.161262 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-combined-ca-bundle\") pod \"barbican-api-85664c95cd-7bnn9\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.161615 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.184831 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.243481 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69b9b656dc-jwxhg" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.532320 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="ceilometer-central-agent" containerID="cri-o://86e1161727ab2cd73b072fdf519043d6472e246c487e974f9831da74a28761ca" gracePeriod=30 Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.532716 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerStarted","Data":"5ce6a172cada9e142900e150f3988b104d641127aa700ee591498007b95ae4d1"} Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.544721 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.533106 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="sg-core" containerID="cri-o://1a09894c33583101467d2d5dba0247e78fc51e16cbb109ee8485a72316353cea" gracePeriod=30 Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.533563 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="ceilometer-notification-agent" containerID="cri-o://e8b5d8776713807502178db7368884a6b93fe4c2c627a0b5fedec2752f76bab6" gracePeriod=30 Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.533023 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="proxy-httpd" containerID="cri-o://5ce6a172cada9e142900e150f3988b104d641127aa700ee591498007b95ae4d1" gracePeriod=30 Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.585830 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.065952286 podStartE2EDuration="55.585803084s" podCreationTimestamp="2025-11-22 08:24:07 +0000 UTC" firstStartedPulling="2025-11-22 08:24:09.584590816 +0000 UTC m=+1271.188929421" lastFinishedPulling="2025-11-22 08:25:01.104441614 +0000 UTC m=+1322.708780219" observedRunningTime="2025-11-22 08:25:02.563088192 +0000 UTC m=+1324.167426817" watchObservedRunningTime="2025-11-22 08:25:02.585803084 +0000 UTC m=+1324.190141699" Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.683208 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-66b55c68fd-lfj8p"] Nov 22 08:25:02 crc kubenswrapper[4735]: I1122 08:25:02.870958 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-76v9z"] Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.188411 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85664c95cd-7bnn9"] Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.205682 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69b9b656dc-jwxhg"] Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.279998 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd32522c-f1ea-4f33-9e30-7c574e3b427d" path="/var/lib/kubelet/pods/bd32522c-f1ea-4f33-9e30-7c574e3b427d/volumes" Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.543247 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69b9b656dc-jwxhg" event={"ID":"0966ff10-5a13-45a4-a20a-5eb889270c5b","Type":"ContainerStarted","Data":"b7e20ab88ce73f0e246f6de467a2780589881cf617bb864b02d0ac38eee87a29"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.545889 4735 generic.go:334] "Generic (PLEG): container finished" podID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" containerID="02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94" exitCode=0 Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.545981 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" event={"ID":"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b","Type":"ContainerDied","Data":"02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.546011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" event={"ID":"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b","Type":"ContainerStarted","Data":"7651859aa3c89224d14ec28340a56467abf26972086ff3bb1c91c0a72b0bf85d"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.551281 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85664c95cd-7bnn9" event={"ID":"a80806be-71c0-442d-95df-a8183640736b","Type":"ContainerStarted","Data":"ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.551333 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85664c95cd-7bnn9" event={"ID":"a80806be-71c0-442d-95df-a8183640736b","Type":"ContainerStarted","Data":"fb60145c124877b2ecae9efb9262361c23ccd64f580e3f9411dd95500807a38b"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.560028 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" event={"ID":"736e8f87-4b05-4335-954a-47d3c89f84fc","Type":"ContainerStarted","Data":"bafaa1481184bb763a5335b271fe80bea9f1cde438c2b683288c9900d658d49c"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.570021 4735 generic.go:334] "Generic (PLEG): container finished" podID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerID="5ce6a172cada9e142900e150f3988b104d641127aa700ee591498007b95ae4d1" exitCode=0 Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.570063 4735 generic.go:334] "Generic (PLEG): container finished" podID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerID="1a09894c33583101467d2d5dba0247e78fc51e16cbb109ee8485a72316353cea" exitCode=2 Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.570071 4735 generic.go:334] "Generic (PLEG): container finished" podID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerID="86e1161727ab2cd73b072fdf519043d6472e246c487e974f9831da74a28761ca" exitCode=0 Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.570136 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerDied","Data":"5ce6a172cada9e142900e150f3988b104d641127aa700ee591498007b95ae4d1"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.570166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerDied","Data":"1a09894c33583101467d2d5dba0247e78fc51e16cbb109ee8485a72316353cea"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.570177 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerDied","Data":"86e1161727ab2cd73b072fdf519043d6472e246c487e974f9831da74a28761ca"} Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.572714 4735 generic.go:334] "Generic (PLEG): container finished" podID="c9f75ef9-02e1-4af0-b532-2e744d02d842" containerID="6a111f241484eb5391838b9af58a78e8bf22389fb19a1d2438956734070d506d" exitCode=0 Nov 22 08:25:03 crc kubenswrapper[4735]: I1122 08:25:03.572894 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qsp4r" event={"ID":"c9f75ef9-02e1-4af0-b532-2e744d02d842","Type":"ContainerDied","Data":"6a111f241484eb5391838b9af58a78e8bf22389fb19a1d2438956734070d506d"} Nov 22 08:25:04 crc kubenswrapper[4735]: I1122 08:25:04.632193 4735 generic.go:334] "Generic (PLEG): container finished" podID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerID="e8b5d8776713807502178db7368884a6b93fe4c2c627a0b5fedec2752f76bab6" exitCode=0 Nov 22 08:25:04 crc kubenswrapper[4735]: I1122 08:25:04.632695 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerDied","Data":"e8b5d8776713807502178db7368884a6b93fe4c2c627a0b5fedec2752f76bab6"} Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.119235 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-cdffccfd-kxg89"] Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.121332 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.124122 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.124132 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.136238 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cdffccfd-kxg89"] Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.246500 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-internal-tls-certs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.246681 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-config-data-custom\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.246745 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-combined-ca-bundle\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.246962 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-config-data\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.247012 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj7rl\" (UniqueName: \"kubernetes.io/projected/9993180a-9f49-4fce-a831-ca0b4ef58428-kube-api-access-gj7rl\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.247245 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-public-tls-certs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.247321 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9993180a-9f49-4fce-a831-ca0b4ef58428-logs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.348583 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9993180a-9f49-4fce-a831-ca0b4ef58428-logs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.348638 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-internal-tls-certs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.348683 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-config-data-custom\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.348704 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-combined-ca-bundle\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.348812 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-config-data\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.348834 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj7rl\" (UniqueName: \"kubernetes.io/projected/9993180a-9f49-4fce-a831-ca0b4ef58428-kube-api-access-gj7rl\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.348920 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-public-tls-certs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.350160 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9993180a-9f49-4fce-a831-ca0b4ef58428-logs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.355770 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-public-tls-certs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.358783 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-combined-ca-bundle\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.363875 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-config-data\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.372987 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj7rl\" (UniqueName: \"kubernetes.io/projected/9993180a-9f49-4fce-a831-ca0b4ef58428-kube-api-access-gj7rl\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.378742 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-internal-tls-certs\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.379926 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9993180a-9f49-4fce-a831-ca0b4ef58428-config-data-custom\") pod \"barbican-api-cdffccfd-kxg89\" (UID: \"9993180a-9f49-4fce-a831-ca0b4ef58428\") " pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.446783 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.659901 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85664c95cd-7bnn9" event={"ID":"a80806be-71c0-442d-95df-a8183640736b","Type":"ContainerStarted","Data":"0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a"} Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.660279 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.665626 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a7f4352-7c85-4d46-b0f6-f851fd011d2b","Type":"ContainerDied","Data":"26c2b5f297b401d3866e16b2cb1be602e30f09f9d824bbeb0beab395d956a791"} Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.665669 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26c2b5f297b401d3866e16b2cb1be602e30f09f9d824bbeb0beab395d956a791" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.667474 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qsp4r" event={"ID":"c9f75ef9-02e1-4af0-b532-2e744d02d842","Type":"ContainerDied","Data":"3a2c6fa0b1cc36f0dd3c29b222c23070d1b5e5870b62e5953d1fb13555a62e62"} Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.667497 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a2c6fa0b1cc36f0dd3c29b222c23070d1b5e5870b62e5953d1fb13555a62e62" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.684216 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-85664c95cd-7bnn9" podStartSLOduration=4.684194107 podStartE2EDuration="4.684194107s" podCreationTimestamp="2025-11-22 08:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:05.682641623 +0000 UTC m=+1327.286980228" watchObservedRunningTime="2025-11-22 08:25:05.684194107 +0000 UTC m=+1327.288532702" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.708669 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.734338 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859501 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-scripts\") pod \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859597 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-sg-core-conf-yaml\") pod \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859676 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-db-sync-config-data\") pod \"c9f75ef9-02e1-4af0-b532-2e744d02d842\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859724 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smh5b\" (UniqueName: \"kubernetes.io/projected/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-kube-api-access-smh5b\") pod \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859752 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-config-data\") pod \"c9f75ef9-02e1-4af0-b532-2e744d02d842\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859782 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c9f75ef9-02e1-4af0-b532-2e744d02d842-etc-machine-id\") pod \"c9f75ef9-02e1-4af0-b532-2e744d02d842\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859852 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-scripts\") pod \"c9f75ef9-02e1-4af0-b532-2e744d02d842\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859969 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqjnv\" (UniqueName: \"kubernetes.io/projected/c9f75ef9-02e1-4af0-b532-2e744d02d842-kube-api-access-tqjnv\") pod \"c9f75ef9-02e1-4af0-b532-2e744d02d842\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.859975 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9f75ef9-02e1-4af0-b532-2e744d02d842-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c9f75ef9-02e1-4af0-b532-2e744d02d842" (UID: "c9f75ef9-02e1-4af0-b532-2e744d02d842"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.860023 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-combined-ca-bundle\") pod \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.860131 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-log-httpd\") pod \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.860247 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-run-httpd\") pod \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.860279 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-config-data\") pod \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\" (UID: \"2a7f4352-7c85-4d46-b0f6-f851fd011d2b\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.860484 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-combined-ca-bundle\") pod \"c9f75ef9-02e1-4af0-b532-2e744d02d842\" (UID: \"c9f75ef9-02e1-4af0-b532-2e744d02d842\") " Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.861118 4735 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c9f75ef9-02e1-4af0-b532-2e744d02d842-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.861255 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2a7f4352-7c85-4d46-b0f6-f851fd011d2b" (UID: "2a7f4352-7c85-4d46-b0f6-f851fd011d2b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.861614 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2a7f4352-7c85-4d46-b0f6-f851fd011d2b" (UID: "2a7f4352-7c85-4d46-b0f6-f851fd011d2b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.863523 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-scripts" (OuterVolumeSpecName: "scripts") pod "c9f75ef9-02e1-4af0-b532-2e744d02d842" (UID: "c9f75ef9-02e1-4af0-b532-2e744d02d842"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.865637 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-scripts" (OuterVolumeSpecName: "scripts") pod "2a7f4352-7c85-4d46-b0f6-f851fd011d2b" (UID: "2a7f4352-7c85-4d46-b0f6-f851fd011d2b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.865805 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9f75ef9-02e1-4af0-b532-2e744d02d842-kube-api-access-tqjnv" (OuterVolumeSpecName: "kube-api-access-tqjnv") pod "c9f75ef9-02e1-4af0-b532-2e744d02d842" (UID: "c9f75ef9-02e1-4af0-b532-2e744d02d842"). InnerVolumeSpecName "kube-api-access-tqjnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.868626 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-kube-api-access-smh5b" (OuterVolumeSpecName: "kube-api-access-smh5b") pod "2a7f4352-7c85-4d46-b0f6-f851fd011d2b" (UID: "2a7f4352-7c85-4d46-b0f6-f851fd011d2b"). InnerVolumeSpecName "kube-api-access-smh5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.869616 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c9f75ef9-02e1-4af0-b532-2e744d02d842" (UID: "c9f75ef9-02e1-4af0-b532-2e744d02d842"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.890281 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2a7f4352-7c85-4d46-b0f6-f851fd011d2b" (UID: "2a7f4352-7c85-4d46-b0f6-f851fd011d2b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.904138 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9f75ef9-02e1-4af0-b532-2e744d02d842" (UID: "c9f75ef9-02e1-4af0-b532-2e744d02d842"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.933747 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-config-data" (OuterVolumeSpecName: "config-data") pod "c9f75ef9-02e1-4af0-b532-2e744d02d842" (UID: "c9f75ef9-02e1-4af0-b532-2e744d02d842"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.936278 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cdffccfd-kxg89"] Nov 22 08:25:05 crc kubenswrapper[4735]: W1122 08:25:05.937408 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9993180a_9f49_4fce_a831_ca0b4ef58428.slice/crio-eed10239206455813f7e76067ae8051fd6688a8902b1917b398edb6b7b98e984 WatchSource:0}: Error finding container eed10239206455813f7e76067ae8051fd6688a8902b1917b398edb6b7b98e984: Status 404 returned error can't find the container with id eed10239206455813f7e76067ae8051fd6688a8902b1917b398edb6b7b98e984 Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.963722 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqjnv\" (UniqueName: \"kubernetes.io/projected/c9f75ef9-02e1-4af0-b532-2e744d02d842-kube-api-access-tqjnv\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.963755 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.963764 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.963774 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.963782 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.963790 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.964084 4735 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.964099 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.964108 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smh5b\" (UniqueName: \"kubernetes.io/projected/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-kube-api-access-smh5b\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.964116 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9f75ef9-02e1-4af0-b532-2e744d02d842-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:05 crc kubenswrapper[4735]: I1122 08:25:05.997271 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-config-data" (OuterVolumeSpecName: "config-data") pod "2a7f4352-7c85-4d46-b0f6-f851fd011d2b" (UID: "2a7f4352-7c85-4d46-b0f6-f851fd011d2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.041017 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a7f4352-7c85-4d46-b0f6-f851fd011d2b" (UID: "2a7f4352-7c85-4d46-b0f6-f851fd011d2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.065815 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.065858 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a7f4352-7c85-4d46-b0f6-f851fd011d2b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.681600 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" event={"ID":"736e8f87-4b05-4335-954a-47d3c89f84fc","Type":"ContainerStarted","Data":"f26f5eae6e43d07447642f89dc0c0b47bebaf802c7dd1d514ce71b2041ce791a"} Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.681925 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" event={"ID":"736e8f87-4b05-4335-954a-47d3c89f84fc","Type":"ContainerStarted","Data":"3cd2c59028197766d287258d5cf88fae793f1b11c1f6780258cab42ba147c5f9"} Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.686725 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cdffccfd-kxg89" event={"ID":"9993180a-9f49-4fce-a831-ca0b4ef58428","Type":"ContainerStarted","Data":"3f26491272f0aa3ef30cf2879f26fdfce74738e9235903cb04e91975e9b85e0f"} Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.686769 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cdffccfd-kxg89" event={"ID":"9993180a-9f49-4fce-a831-ca0b4ef58428","Type":"ContainerStarted","Data":"154196eb1aa2c283a8f332863370e40ea1d70231a6ffbee37c53d5e485deac2c"} Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.686780 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cdffccfd-kxg89" event={"ID":"9993180a-9f49-4fce-a831-ca0b4ef58428","Type":"ContainerStarted","Data":"eed10239206455813f7e76067ae8051fd6688a8902b1917b398edb6b7b98e984"} Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.686849 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.689126 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69b9b656dc-jwxhg" event={"ID":"0966ff10-5a13-45a4-a20a-5eb889270c5b","Type":"ContainerStarted","Data":"46e84cdb4d788f5c89daf5a4613a9d47b2978bbd6bb165eee8ddb03b3ceb8427"} Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.689184 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69b9b656dc-jwxhg" event={"ID":"0966ff10-5a13-45a4-a20a-5eb889270c5b","Type":"ContainerStarted","Data":"43882b5c7d783b9ae846fada1a81cac3f31d947c8fa81c8b301ae65ba363cc6b"} Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.692087 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" event={"ID":"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b","Type":"ContainerStarted","Data":"82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc"} Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.692123 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qsp4r" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.692205 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.692652 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.708065 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-66b55c68fd-lfj8p" podStartSLOduration=3.987602022 podStartE2EDuration="5.708046722s" podCreationTimestamp="2025-11-22 08:25:01 +0000 UTC" firstStartedPulling="2025-11-22 08:25:02.716490808 +0000 UTC m=+1324.320829413" lastFinishedPulling="2025-11-22 08:25:04.436935508 +0000 UTC m=+1326.041274113" observedRunningTime="2025-11-22 08:25:06.70075843 +0000 UTC m=+1328.305097035" watchObservedRunningTime="2025-11-22 08:25:06.708046722 +0000 UTC m=+1328.312385327" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.755298 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" podStartSLOduration=5.755282806 podStartE2EDuration="5.755282806s" podCreationTimestamp="2025-11-22 08:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:06.754874784 +0000 UTC m=+1328.359213389" watchObservedRunningTime="2025-11-22 08:25:06.755282806 +0000 UTC m=+1328.359621411" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.761864 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-69b9b656dc-jwxhg" podStartSLOduration=3.435862808 podStartE2EDuration="5.761847218s" podCreationTimestamp="2025-11-22 08:25:01 +0000 UTC" firstStartedPulling="2025-11-22 08:25:03.174719013 +0000 UTC m=+1324.779057608" lastFinishedPulling="2025-11-22 08:25:05.500703413 +0000 UTC m=+1327.105042018" observedRunningTime="2025-11-22 08:25:06.726335331 +0000 UTC m=+1328.330673936" watchObservedRunningTime="2025-11-22 08:25:06.761847218 +0000 UTC m=+1328.366185823" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.795221 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-cdffccfd-kxg89" podStartSLOduration=1.795205856 podStartE2EDuration="1.795205856s" podCreationTimestamp="2025-11-22 08:25:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:06.782858753 +0000 UTC m=+1328.387197358" watchObservedRunningTime="2025-11-22 08:25:06.795205856 +0000 UTC m=+1328.399544461" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.822119 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.851033 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.884835 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:06 crc kubenswrapper[4735]: E1122 08:25:06.886715 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="sg-core" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.886741 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="sg-core" Nov 22 08:25:06 crc kubenswrapper[4735]: E1122 08:25:06.886762 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="ceilometer-notification-agent" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.886769 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="ceilometer-notification-agent" Nov 22 08:25:06 crc kubenswrapper[4735]: E1122 08:25:06.886798 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="proxy-httpd" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.886805 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="proxy-httpd" Nov 22 08:25:06 crc kubenswrapper[4735]: E1122 08:25:06.886829 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f75ef9-02e1-4af0-b532-2e744d02d842" containerName="cinder-db-sync" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.886835 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f75ef9-02e1-4af0-b532-2e744d02d842" containerName="cinder-db-sync" Nov 22 08:25:06 crc kubenswrapper[4735]: E1122 08:25:06.886881 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="ceilometer-central-agent" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.886887 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="ceilometer-central-agent" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.891336 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="sg-core" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.891421 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="ceilometer-central-agent" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.891453 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f75ef9-02e1-4af0-b532-2e744d02d842" containerName="cinder-db-sync" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.891489 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="proxy-httpd" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.891509 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" containerName="ceilometer-notification-agent" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.895904 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.918850 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.919137 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:25:06 crc kubenswrapper[4735]: I1122 08:25:06.929832 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.023844 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-scripts\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.023970 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-run-httpd\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.024002 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fgpg\" (UniqueName: \"kubernetes.io/projected/a8d81c52-8a16-40a1-b2c2-1a476275b92c-kube-api-access-4fgpg\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.024065 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.024140 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-log-httpd\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.024206 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-config-data\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.024242 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.067309 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.069370 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.074324 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.074557 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-chn6w" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.074697 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.074866 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.099265 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.129615 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.129659 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/486b072b-b195-440f-9d34-62a962bcbba6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.129704 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-log-httpd\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.129811 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-config-data\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.129886 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.129945 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-scripts\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.129981 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-scripts\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.130022 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.130207 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-run-httpd\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.130250 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fgpg\" (UniqueName: \"kubernetes.io/projected/a8d81c52-8a16-40a1-b2c2-1a476275b92c-kube-api-access-4fgpg\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.130277 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.130322 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7z24\" (UniqueName: \"kubernetes.io/projected/486b072b-b195-440f-9d34-62a962bcbba6-kube-api-access-w7z24\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.130365 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.132423 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-log-httpd\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.132767 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-run-httpd\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.147134 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-config-data\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.150881 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-scripts\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.182126 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.183254 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.188860 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.231997 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.232181 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7z24\" (UniqueName: \"kubernetes.io/projected/486b072b-b195-440f-9d34-62a962bcbba6-kube-api-access-w7z24\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.243180 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.251032 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/486b072b-b195-440f-9d34-62a962bcbba6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.251529 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-scripts\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.251643 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.252153 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.252236 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/486b072b-b195-440f-9d34-62a962bcbba6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.241786 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.233331 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fgpg\" (UniqueName: \"kubernetes.io/projected/a8d81c52-8a16-40a1-b2c2-1a476275b92c-kube-api-access-4fgpg\") pod \"ceilometer-0\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.258948 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-scripts\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.278799 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.281060 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.287154 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7z24\" (UniqueName: \"kubernetes.io/projected/486b072b-b195-440f-9d34-62a962bcbba6-kube-api-access-w7z24\") pod \"cinder-scheduler-0\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.302690 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a7f4352-7c85-4d46-b0f6-f851fd011d2b" path="/var/lib/kubelet/pods/2a7f4352-7c85-4d46-b0f6-f851fd011d2b/volumes" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.303711 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-76v9z"] Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.303737 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5f5xm"] Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.308832 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.349954 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5f5xm"] Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.353712 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r5m9\" (UniqueName: \"kubernetes.io/projected/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-kube-api-access-7r5m9\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.353790 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.353830 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.353897 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.353979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.354006 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-config\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.394792 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.401489 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.401657 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.407601 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.451526 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.456768 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.456843 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.456876 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-config\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.456894 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704f5bd0-e968-47c0-a605-0f0d56023e81-logs\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.456946 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r5m9\" (UniqueName: \"kubernetes.io/projected/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-kube-api-access-7r5m9\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.456977 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/704f5bd0-e968-47c0-a605-0f0d56023e81-etc-machine-id\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.457002 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-scripts\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.457020 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.457058 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.457076 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.457120 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.457142 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data-custom\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.457172 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk9xk\" (UniqueName: \"kubernetes.io/projected/704f5bd0-e968-47c0-a605-0f0d56023e81-kube-api-access-jk9xk\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.463068 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-config\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.463312 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.465211 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.466163 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.469415 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.482115 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r5m9\" (UniqueName: \"kubernetes.io/projected/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-kube-api-access-7r5m9\") pod \"dnsmasq-dns-5c9776ccc5-5f5xm\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.569437 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.569653 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data-custom\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.569726 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk9xk\" (UniqueName: \"kubernetes.io/projected/704f5bd0-e968-47c0-a605-0f0d56023e81-kube-api-access-jk9xk\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.569806 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.570284 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704f5bd0-e968-47c0-a605-0f0d56023e81-logs\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.570420 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/704f5bd0-e968-47c0-a605-0f0d56023e81-etc-machine-id\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.570502 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-scripts\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.575906 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704f5bd0-e968-47c0-a605-0f0d56023e81-logs\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.577193 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data-custom\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.577365 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/704f5bd0-e968-47c0-a605-0f0d56023e81-etc-machine-id\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.589503 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.590655 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-scripts\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.596056 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.602006 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk9xk\" (UniqueName: \"kubernetes.io/projected/704f5bd0-e968-47c0-a605-0f0d56023e81-kube-api-access-jk9xk\") pod \"cinder-api-0\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.691079 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.709159 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:25:07 crc kubenswrapper[4735]: I1122 08:25:07.715714 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.072686 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:08 crc kubenswrapper[4735]: W1122 08:25:08.234585 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod486b072b_b195_440f_9d34_62a962bcbba6.slice/crio-9f3973eb1ab41ea061eba1944acb4acf61ad2297ae737bfe757f0f3d810818f4 WatchSource:0}: Error finding container 9f3973eb1ab41ea061eba1944acb4acf61ad2297ae737bfe757f0f3d810818f4: Status 404 returned error can't find the container with id 9f3973eb1ab41ea061eba1944acb4acf61ad2297ae737bfe757f0f3d810818f4 Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.243137 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.370873 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5f5xm"] Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.447670 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.723815 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"704f5bd0-e968-47c0-a605-0f0d56023e81","Type":"ContainerStarted","Data":"aca4109a644688133a7d33f6c652440316c10b14eff8ec6eabb6ab93c3776b1b"} Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.724947 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"486b072b-b195-440f-9d34-62a962bcbba6","Type":"ContainerStarted","Data":"9f3973eb1ab41ea061eba1944acb4acf61ad2297ae737bfe757f0f3d810818f4"} Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.725972 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" event={"ID":"f73d0ea1-05cf-413c-b5f8-2bef40df95b8","Type":"ContainerStarted","Data":"cf36dacc03afd6d1451b24e76e6be33a4764831dcde1c654f45d8c5ab8a18093"} Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.726885 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerStarted","Data":"6a6cf54ba44c2d49464e7119d2b76890e1e3e0378f82ef10b37a29fc616779fa"} Nov 22 08:25:08 crc kubenswrapper[4735]: I1122 08:25:08.727114 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" podUID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" containerName="dnsmasq-dns" containerID="cri-o://82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc" gracePeriod=10 Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.260242 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.305660 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.324813 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsxs4\" (UniqueName: \"kubernetes.io/projected/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-kube-api-access-rsxs4\") pod \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.325361 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-nb\") pod \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.325404 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-config\") pod \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.325443 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-swift-storage-0\") pod \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.325587 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-sb\") pod \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.325627 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-svc\") pod \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\" (UID: \"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b\") " Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.332631 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-kube-api-access-rsxs4" (OuterVolumeSpecName: "kube-api-access-rsxs4") pod "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" (UID: "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b"). InnerVolumeSpecName "kube-api-access-rsxs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.428048 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsxs4\" (UniqueName: \"kubernetes.io/projected/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-kube-api-access-rsxs4\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.490228 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" (UID: "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.536988 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.695151 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" (UID: "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.709178 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" (UID: "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.729028 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-config" (OuterVolumeSpecName: "config") pod "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" (UID: "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.745604 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.745641 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.745654 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.790281 4735 generic.go:334] "Generic (PLEG): container finished" podID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" containerID="82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc" exitCode=0 Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.790346 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" event={"ID":"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b","Type":"ContainerDied","Data":"82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc"} Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.790374 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" event={"ID":"ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b","Type":"ContainerDied","Data":"7651859aa3c89224d14ec28340a56467abf26972086ff3bb1c91c0a72b0bf85d"} Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.790390 4735 scope.go:117] "RemoveContainer" containerID="82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.790533 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-76v9z" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.802073 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"704f5bd0-e968-47c0-a605-0f0d56023e81","Type":"ContainerStarted","Data":"12e61705ef784d5c0396940a6a3476abc45478d982f44322ba112446fc102109"} Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.807956 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" (UID: "ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.808194 4735 generic.go:334] "Generic (PLEG): container finished" podID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" containerID="c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150" exitCode=0 Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.808286 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" event={"ID":"f73d0ea1-05cf-413c-b5f8-2bef40df95b8","Type":"ContainerDied","Data":"c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150"} Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.822039 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerStarted","Data":"8ebb0c9292dcb1041f95e192c1b40e3b7a324dd2bb34acddf3dd4a05bf1feb2b"} Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.824871 4735 scope.go:117] "RemoveContainer" containerID="02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.848444 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.896372 4735 scope.go:117] "RemoveContainer" containerID="82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc" Nov 22 08:25:09 crc kubenswrapper[4735]: E1122 08:25:09.898023 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc\": container with ID starting with 82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc not found: ID does not exist" containerID="82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.898059 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc"} err="failed to get container status \"82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc\": rpc error: code = NotFound desc = could not find container \"82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc\": container with ID starting with 82f84ac0fabe559bc7924d3fe10bb8caa198482c1725b1814cfd6e50ffdbc5cc not found: ID does not exist" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.898099 4735 scope.go:117] "RemoveContainer" containerID="02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94" Nov 22 08:25:09 crc kubenswrapper[4735]: E1122 08:25:09.898385 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94\": container with ID starting with 02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94 not found: ID does not exist" containerID="02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94" Nov 22 08:25:09 crc kubenswrapper[4735]: I1122 08:25:09.898432 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94"} err="failed to get container status \"02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94\": rpc error: code = NotFound desc = could not find container \"02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94\": container with ID starting with 02f0a77448f037df00aff1913567b374bb4fc5512a5381dc599f1d971f8dea94 not found: ID does not exist" Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.170394 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-76v9z"] Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.196340 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-76v9z"] Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.854920 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerStarted","Data":"c149bb79a9740082c76c781ede88172aff86af44fbfbd02d97318e12f40634b1"} Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.855326 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerStarted","Data":"cbcb10f4e1ead4c3a62dcdb5ecc619d242a4847559407a08570669f7b6c82553"} Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.868417 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"704f5bd0-e968-47c0-a605-0f0d56023e81","Type":"ContainerStarted","Data":"76f50ffbaf53f10a808c31db0ca8c755187aee83b8aff0683d0c3c4a9382f667"} Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.868651 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerName="cinder-api-log" containerID="cri-o://12e61705ef784d5c0396940a6a3476abc45478d982f44322ba112446fc102109" gracePeriod=30 Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.868730 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.869121 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerName="cinder-api" containerID="cri-o://76f50ffbaf53f10a808c31db0ca8c755187aee83b8aff0683d0c3c4a9382f667" gracePeriod=30 Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.878087 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"486b072b-b195-440f-9d34-62a962bcbba6","Type":"ContainerStarted","Data":"8f7c6544be271f5b2aae181961fe1de597cb18b3210467d04623bbdfced54728"} Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.888330 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.888308953 podStartE2EDuration="3.888308953s" podCreationTimestamp="2025-11-22 08:25:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:10.883282034 +0000 UTC m=+1332.487620639" watchObservedRunningTime="2025-11-22 08:25:10.888308953 +0000 UTC m=+1332.492647558" Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.903740 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" event={"ID":"f73d0ea1-05cf-413c-b5f8-2bef40df95b8","Type":"ContainerStarted","Data":"9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a"} Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.904719 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:10 crc kubenswrapper[4735]: I1122 08:25:10.932358 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" podStartSLOduration=3.932338858 podStartE2EDuration="3.932338858s" podCreationTimestamp="2025-11-22 08:25:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:10.921134496 +0000 UTC m=+1332.525473101" watchObservedRunningTime="2025-11-22 08:25:10.932338858 +0000 UTC m=+1332.536677463" Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.284323 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" path="/var/lib/kubelet/pods/ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b/volumes" Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.918147 4735 generic.go:334] "Generic (PLEG): container finished" podID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerID="76f50ffbaf53f10a808c31db0ca8c755187aee83b8aff0683d0c3c4a9382f667" exitCode=0 Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.918400 4735 generic.go:334] "Generic (PLEG): container finished" podID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerID="12e61705ef784d5c0396940a6a3476abc45478d982f44322ba112446fc102109" exitCode=143 Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.918335 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"704f5bd0-e968-47c0-a605-0f0d56023e81","Type":"ContainerDied","Data":"76f50ffbaf53f10a808c31db0ca8c755187aee83b8aff0683d0c3c4a9382f667"} Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.918481 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"704f5bd0-e968-47c0-a605-0f0d56023e81","Type":"ContainerDied","Data":"12e61705ef784d5c0396940a6a3476abc45478d982f44322ba112446fc102109"} Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.918496 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"704f5bd0-e968-47c0-a605-0f0d56023e81","Type":"ContainerDied","Data":"aca4109a644688133a7d33f6c652440316c10b14eff8ec6eabb6ab93c3776b1b"} Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.918506 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aca4109a644688133a7d33f6c652440316c10b14eff8ec6eabb6ab93c3776b1b" Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.927541 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"486b072b-b195-440f-9d34-62a962bcbba6","Type":"ContainerStarted","Data":"43c2e6cad4d45b375688baf10460e72b6ddb138c67e3a13d62b310e381e3787c"} Nov 22 08:25:11 crc kubenswrapper[4735]: I1122 08:25:11.958857 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.118377553 podStartE2EDuration="4.958830938s" podCreationTimestamp="2025-11-22 08:25:07 +0000 UTC" firstStartedPulling="2025-11-22 08:25:08.241394548 +0000 UTC m=+1329.845733153" lastFinishedPulling="2025-11-22 08:25:09.081847933 +0000 UTC m=+1330.686186538" observedRunningTime="2025-11-22 08:25:11.952630374 +0000 UTC m=+1333.556968979" watchObservedRunningTime="2025-11-22 08:25:11.958830938 +0000 UTC m=+1333.563169543" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.010921 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.096036 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk9xk\" (UniqueName: \"kubernetes.io/projected/704f5bd0-e968-47c0-a605-0f0d56023e81-kube-api-access-jk9xk\") pod \"704f5bd0-e968-47c0-a605-0f0d56023e81\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.096262 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data-custom\") pod \"704f5bd0-e968-47c0-a605-0f0d56023e81\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.096844 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-combined-ca-bundle\") pod \"704f5bd0-e968-47c0-a605-0f0d56023e81\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.096892 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704f5bd0-e968-47c0-a605-0f0d56023e81-logs\") pod \"704f5bd0-e968-47c0-a605-0f0d56023e81\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.096951 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-scripts\") pod \"704f5bd0-e968-47c0-a605-0f0d56023e81\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.097056 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data\") pod \"704f5bd0-e968-47c0-a605-0f0d56023e81\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.097103 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/704f5bd0-e968-47c0-a605-0f0d56023e81-etc-machine-id\") pod \"704f5bd0-e968-47c0-a605-0f0d56023e81\" (UID: \"704f5bd0-e968-47c0-a605-0f0d56023e81\") " Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.102499 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/704f5bd0-e968-47c0-a605-0f0d56023e81-logs" (OuterVolumeSpecName: "logs") pod "704f5bd0-e968-47c0-a605-0f0d56023e81" (UID: "704f5bd0-e968-47c0-a605-0f0d56023e81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.102586 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/704f5bd0-e968-47c0-a605-0f0d56023e81-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "704f5bd0-e968-47c0-a605-0f0d56023e81" (UID: "704f5bd0-e968-47c0-a605-0f0d56023e81"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.111638 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "704f5bd0-e968-47c0-a605-0f0d56023e81" (UID: "704f5bd0-e968-47c0-a605-0f0d56023e81"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.113768 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/704f5bd0-e968-47c0-a605-0f0d56023e81-kube-api-access-jk9xk" (OuterVolumeSpecName: "kube-api-access-jk9xk") pod "704f5bd0-e968-47c0-a605-0f0d56023e81" (UID: "704f5bd0-e968-47c0-a605-0f0d56023e81"). InnerVolumeSpecName "kube-api-access-jk9xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.114193 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-scripts" (OuterVolumeSpecName: "scripts") pod "704f5bd0-e968-47c0-a605-0f0d56023e81" (UID: "704f5bd0-e968-47c0-a605-0f0d56023e81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.153637 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "704f5bd0-e968-47c0-a605-0f0d56023e81" (UID: "704f5bd0-e968-47c0-a605-0f0d56023e81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.199916 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.199952 4735 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/704f5bd0-e968-47c0-a605-0f0d56023e81-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.199966 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk9xk\" (UniqueName: \"kubernetes.io/projected/704f5bd0-e968-47c0-a605-0f0d56023e81-kube-api-access-jk9xk\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.199974 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.199982 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.199991 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704f5bd0-e968-47c0-a605-0f0d56023e81-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.206967 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data" (OuterVolumeSpecName: "config-data") pod "704f5bd0-e968-47c0-a605-0f0d56023e81" (UID: "704f5bd0-e968-47c0-a605-0f0d56023e81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.301784 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704f5bd0-e968-47c0-a605-0f0d56023e81-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.402398 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.952675 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerStarted","Data":"c8ea1500e428226bdd507961f97a1f2fc9ed4a47730c05233c5f59f53d35e882"} Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.952694 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.952980 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:25:12 crc kubenswrapper[4735]: I1122 08:25:12.988908 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.240979468 podStartE2EDuration="6.988889515s" podCreationTimestamp="2025-11-22 08:25:06 +0000 UTC" firstStartedPulling="2025-11-22 08:25:08.09619977 +0000 UTC m=+1329.700538375" lastFinishedPulling="2025-11-22 08:25:11.844109817 +0000 UTC m=+1333.448448422" observedRunningTime="2025-11-22 08:25:12.984412731 +0000 UTC m=+1334.588751346" watchObservedRunningTime="2025-11-22 08:25:12.988889515 +0000 UTC m=+1334.593228120" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.016728 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.032492 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.045924 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:13 crc kubenswrapper[4735]: E1122 08:25:13.046434 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" containerName="dnsmasq-dns" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.046467 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" containerName="dnsmasq-dns" Nov 22 08:25:13 crc kubenswrapper[4735]: E1122 08:25:13.046485 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerName="cinder-api" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.046493 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerName="cinder-api" Nov 22 08:25:13 crc kubenswrapper[4735]: E1122 08:25:13.046504 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" containerName="init" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.046512 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" containerName="init" Nov 22 08:25:13 crc kubenswrapper[4735]: E1122 08:25:13.046524 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerName="cinder-api-log" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.046529 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerName="cinder-api-log" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.046764 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec9cfaaf-aa87-41d0-9cb9-dba59a144c1b" containerName="dnsmasq-dns" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.046786 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerName="cinder-api" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.046804 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" containerName="cinder-api-log" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.053618 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.056061 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.056632 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.056835 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.071728 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.222979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5494e4ea-bbac-4c48-ac26-7141cc543bb4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.223017 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.223037 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5494e4ea-bbac-4c48-ac26-7141cc543bb4-logs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.223181 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-config-data\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.223355 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-config-data-custom\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.223499 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-scripts\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.223654 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.223751 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.223904 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5wb7\" (UniqueName: \"kubernetes.io/projected/5494e4ea-bbac-4c48-ac26-7141cc543bb4-kube-api-access-p5wb7\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.277279 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="704f5bd0-e968-47c0-a605-0f0d56023e81" path="/var/lib/kubelet/pods/704f5bd0-e968-47c0-a605-0f0d56023e81/volumes" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326299 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5wb7\" (UniqueName: \"kubernetes.io/projected/5494e4ea-bbac-4c48-ac26-7141cc543bb4-kube-api-access-p5wb7\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326372 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5494e4ea-bbac-4c48-ac26-7141cc543bb4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326396 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326414 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5494e4ea-bbac-4c48-ac26-7141cc543bb4-logs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326455 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-config-data\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326510 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-config-data-custom\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326551 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-scripts\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326593 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.326628 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.327440 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5494e4ea-bbac-4c48-ac26-7141cc543bb4-logs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.327749 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5494e4ea-bbac-4c48-ac26-7141cc543bb4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.333221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.333391 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-config-data\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.334745 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.335301 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-scripts\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.335686 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.352555 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5494e4ea-bbac-4c48-ac26-7141cc543bb4-config-data-custom\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.362972 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5wb7\" (UniqueName: \"kubernetes.io/projected/5494e4ea-bbac-4c48-ac26-7141cc543bb4-kube-api-access-p5wb7\") pod \"cinder-api-0\" (UID: \"5494e4ea-bbac-4c48-ac26-7141cc543bb4\") " pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.372983 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.710560 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.844695 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.883516 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:25:13 crc kubenswrapper[4735]: I1122 08:25:13.971704 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5494e4ea-bbac-4c48-ac26-7141cc543bb4","Type":"ContainerStarted","Data":"d0a417964657a29b8d53f5232d8ad38308b67e2d4dfe50f7bb24ffc4e9a98674"} Nov 22 08:25:15 crc kubenswrapper[4735]: I1122 08:25:14.999857 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5494e4ea-bbac-4c48-ac26-7141cc543bb4","Type":"ContainerStarted","Data":"66c3c6929b8eaecc42dc34d8cd22b1bc5ce85b5464ebd56e8e595b3957875cef"} Nov 22 08:25:16 crc kubenswrapper[4735]: I1122 08:25:16.017509 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5494e4ea-bbac-4c48-ac26-7141cc543bb4","Type":"ContainerStarted","Data":"82988b5fb13f1ff22d35a01f3086213b6c80232a199227e317aac43e86b0549a"} Nov 22 08:25:16 crc kubenswrapper[4735]: I1122 08:25:16.017969 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 22 08:25:16 crc kubenswrapper[4735]: I1122 08:25:16.059669 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.059650219 podStartE2EDuration="3.059650219s" podCreationTimestamp="2025-11-22 08:25:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:16.040042024 +0000 UTC m=+1337.644380659" watchObservedRunningTime="2025-11-22 08:25:16.059650219 +0000 UTC m=+1337.663988824" Nov 22 08:25:16 crc kubenswrapper[4735]: I1122 08:25:16.922677 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:16 crc kubenswrapper[4735]: I1122 08:25:16.936970 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cdffccfd-kxg89" Nov 22 08:25:17 crc kubenswrapper[4735]: I1122 08:25:17.036238 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85664c95cd-7bnn9"] Nov 22 08:25:17 crc kubenswrapper[4735]: I1122 08:25:17.036579 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85664c95cd-7bnn9" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api" containerID="cri-o://0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a" gracePeriod=30 Nov 22 08:25:17 crc kubenswrapper[4735]: I1122 08:25:17.039841 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85664c95cd-7bnn9" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api-log" containerID="cri-o://ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813" gracePeriod=30 Nov 22 08:25:17 crc kubenswrapper[4735]: I1122 08:25:17.647178 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 22 08:25:17 crc kubenswrapper[4735]: I1122 08:25:17.693647 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:17 crc kubenswrapper[4735]: I1122 08:25:17.702501 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:17 crc kubenswrapper[4735]: I1122 08:25:17.772678 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hm5w8"] Nov 22 08:25:17 crc kubenswrapper[4735]: I1122 08:25:17.775152 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" podUID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" containerName="dnsmasq-dns" containerID="cri-o://543c4b61014f49d2ae98f0082d0f4a68d4a1fae8f29f3859fcecaaca5f06da64" gracePeriod=10 Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.073293 4735 generic.go:334] "Generic (PLEG): container finished" podID="a80806be-71c0-442d-95df-a8183640736b" containerID="ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813" exitCode=143 Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.073345 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85664c95cd-7bnn9" event={"ID":"a80806be-71c0-442d-95df-a8183640736b","Type":"ContainerDied","Data":"ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813"} Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.078542 4735 generic.go:334] "Generic (PLEG): container finished" podID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" containerID="543c4b61014f49d2ae98f0082d0f4a68d4a1fae8f29f3859fcecaaca5f06da64" exitCode=0 Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.078699 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" event={"ID":"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded","Type":"ContainerDied","Data":"543c4b61014f49d2ae98f0082d0f4a68d4a1fae8f29f3859fcecaaca5f06da64"} Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.078809 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="486b072b-b195-440f-9d34-62a962bcbba6" containerName="cinder-scheduler" containerID="cri-o://8f7c6544be271f5b2aae181961fe1de597cb18b3210467d04623bbdfced54728" gracePeriod=30 Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.078941 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="486b072b-b195-440f-9d34-62a962bcbba6" containerName="probe" containerID="cri-o://43c2e6cad4d45b375688baf10460e72b6ddb138c67e3a13d62b310e381e3787c" gracePeriod=30 Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.322062 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.442888 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-swift-storage-0\") pod \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.442934 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t9zk\" (UniqueName: \"kubernetes.io/projected/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-kube-api-access-8t9zk\") pod \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.443063 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-sb\") pod \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.443181 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-config\") pod \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.443213 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-svc\") pod \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.443294 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-nb\") pod \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\" (UID: \"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded\") " Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.455745 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-kube-api-access-8t9zk" (OuterVolumeSpecName: "kube-api-access-8t9zk") pod "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" (UID: "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded"). InnerVolumeSpecName "kube-api-access-8t9zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.500154 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-config" (OuterVolumeSpecName: "config") pod "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" (UID: "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.516216 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" (UID: "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.539230 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" (UID: "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.546076 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.546120 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t9zk\" (UniqueName: \"kubernetes.io/projected/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-kube-api-access-8t9zk\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.546109 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" (UID: "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.546134 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.546232 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.546404 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" (UID: "a7e9c91a-b539-4cfd-9b83-3c4e69d9cded"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.648129 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:18 crc kubenswrapper[4735]: I1122 08:25:18.648164 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.090691 4735 generic.go:334] "Generic (PLEG): container finished" podID="486b072b-b195-440f-9d34-62a962bcbba6" containerID="43c2e6cad4d45b375688baf10460e72b6ddb138c67e3a13d62b310e381e3787c" exitCode=0 Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.090761 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"486b072b-b195-440f-9d34-62a962bcbba6","Type":"ContainerDied","Data":"43c2e6cad4d45b375688baf10460e72b6ddb138c67e3a13d62b310e381e3787c"} Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.095371 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" event={"ID":"a7e9c91a-b539-4cfd-9b83-3c4e69d9cded","Type":"ContainerDied","Data":"4099a24b8eefda848208db66f37de6a6daa322fdc076ac7e5aec0157cd4378bf"} Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.095431 4735 scope.go:117] "RemoveContainer" containerID="543c4b61014f49d2ae98f0082d0f4a68d4a1fae8f29f3859fcecaaca5f06da64" Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.095449 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hm5w8" Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.144366 4735 scope.go:117] "RemoveContainer" containerID="1ef143c5226e71c9a1ed59a1fe4ca387120a12a958cef498c700805fc6914400" Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.149713 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hm5w8"] Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.160257 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hm5w8"] Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.341851 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" path="/var/lib/kubelet/pods/a7e9c91a-b539-4cfd-9b83-3c4e69d9cded/volumes" Nov 22 08:25:19 crc kubenswrapper[4735]: I1122 08:25:19.342533 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.247370 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85664c95cd-7bnn9" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.200:9311/healthcheck\": read tcp 10.217.0.2:36086->10.217.0.200:9311: read: connection reset by peer" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.247442 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85664c95cd-7bnn9" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.200:9311/healthcheck\": read tcp 10.217.0.2:36080->10.217.0.200:9311: read: connection reset by peer" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.640915 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fcc565756-vf564" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.687344 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fcc565756-vf564" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.833487 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.919313 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmnfw\" (UniqueName: \"kubernetes.io/projected/a80806be-71c0-442d-95df-a8183640736b-kube-api-access-tmnfw\") pod \"a80806be-71c0-442d-95df-a8183640736b\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.919399 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data\") pod \"a80806be-71c0-442d-95df-a8183640736b\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.919473 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a80806be-71c0-442d-95df-a8183640736b-logs\") pod \"a80806be-71c0-442d-95df-a8183640736b\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.919525 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data-custom\") pod \"a80806be-71c0-442d-95df-a8183640736b\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.919573 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-combined-ca-bundle\") pod \"a80806be-71c0-442d-95df-a8183640736b\" (UID: \"a80806be-71c0-442d-95df-a8183640736b\") " Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.921367 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a80806be-71c0-442d-95df-a8183640736b-logs" (OuterVolumeSpecName: "logs") pod "a80806be-71c0-442d-95df-a8183640736b" (UID: "a80806be-71c0-442d-95df-a8183640736b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.927703 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a80806be-71c0-442d-95df-a8183640736b" (UID: "a80806be-71c0-442d-95df-a8183640736b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.936814 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a80806be-71c0-442d-95df-a8183640736b-kube-api-access-tmnfw" (OuterVolumeSpecName: "kube-api-access-tmnfw") pod "a80806be-71c0-442d-95df-a8183640736b" (UID: "a80806be-71c0-442d-95df-a8183640736b"). InnerVolumeSpecName "kube-api-access-tmnfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:20 crc kubenswrapper[4735]: I1122 08:25:20.958589 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a80806be-71c0-442d-95df-a8183640736b" (UID: "a80806be-71c0-442d-95df-a8183640736b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.001968 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data" (OuterVolumeSpecName: "config-data") pod "a80806be-71c0-442d-95df-a8183640736b" (UID: "a80806be-71c0-442d-95df-a8183640736b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.022573 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.022619 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a80806be-71c0-442d-95df-a8183640736b-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.022631 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.022649 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a80806be-71c0-442d-95df-a8183640736b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.022664 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmnfw\" (UniqueName: \"kubernetes.io/projected/a80806be-71c0-442d-95df-a8183640736b-kube-api-access-tmnfw\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.116158 4735 generic.go:334] "Generic (PLEG): container finished" podID="a80806be-71c0-442d-95df-a8183640736b" containerID="0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a" exitCode=0 Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.116205 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85664c95cd-7bnn9" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.116255 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85664c95cd-7bnn9" event={"ID":"a80806be-71c0-442d-95df-a8183640736b","Type":"ContainerDied","Data":"0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a"} Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.116312 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85664c95cd-7bnn9" event={"ID":"a80806be-71c0-442d-95df-a8183640736b","Type":"ContainerDied","Data":"fb60145c124877b2ecae9efb9262361c23ccd64f580e3f9411dd95500807a38b"} Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.116332 4735 scope.go:117] "RemoveContainer" containerID="0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.152124 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85664c95cd-7bnn9"] Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.154660 4735 scope.go:117] "RemoveContainer" containerID="ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.167926 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-85664c95cd-7bnn9"] Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.181470 4735 scope.go:117] "RemoveContainer" containerID="0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a" Nov 22 08:25:21 crc kubenswrapper[4735]: E1122 08:25:21.181962 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a\": container with ID starting with 0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a not found: ID does not exist" containerID="0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.182028 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a"} err="failed to get container status \"0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a\": rpc error: code = NotFound desc = could not find container \"0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a\": container with ID starting with 0d3cb7e480e65bc277ff8eba218e731859bf0acc0c31cd2651af1815029ccd4a not found: ID does not exist" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.182063 4735 scope.go:117] "RemoveContainer" containerID="ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813" Nov 22 08:25:21 crc kubenswrapper[4735]: E1122 08:25:21.182401 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813\": container with ID starting with ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813 not found: ID does not exist" containerID="ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.182441 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813"} err="failed to get container status \"ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813\": rpc error: code = NotFound desc = could not find container \"ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813\": container with ID starting with ef4973698825575c9152da15702157ed6dbef95304a9c4058442d16454945813 not found: ID does not exist" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.283959 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a80806be-71c0-442d-95df-a8183640736b" path="/var/lib/kubelet/pods/a80806be-71c0-442d-95df-a8183640736b/volumes" Nov 22 08:25:21 crc kubenswrapper[4735]: I1122 08:25:21.569332 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5584fd997b-7ztjv" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.138973 4735 generic.go:334] "Generic (PLEG): container finished" podID="486b072b-b195-440f-9d34-62a962bcbba6" containerID="8f7c6544be271f5b2aae181961fe1de597cb18b3210467d04623bbdfced54728" exitCode=0 Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.139296 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"486b072b-b195-440f-9d34-62a962bcbba6","Type":"ContainerDied","Data":"8f7c6544be271f5b2aae181961fe1de597cb18b3210467d04623bbdfced54728"} Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.185016 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7bb54fbc4c-fnr9g" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.220153 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 22 08:25:22 crc kubenswrapper[4735]: E1122 08:25:22.220798 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.220840 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api" Nov 22 08:25:22 crc kubenswrapper[4735]: E1122 08:25:22.220866 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" containerName="init" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.220876 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" containerName="init" Nov 22 08:25:22 crc kubenswrapper[4735]: E1122 08:25:22.220925 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" containerName="dnsmasq-dns" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.220933 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" containerName="dnsmasq-dns" Nov 22 08:25:22 crc kubenswrapper[4735]: E1122 08:25:22.220946 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api-log" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.220954 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api-log" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.221229 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e9c91a-b539-4cfd-9b83-3c4e69d9cded" containerName="dnsmasq-dns" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.221246 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api-log" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.221277 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a80806be-71c0-442d-95df-a8183640736b" containerName="barbican-api" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.222254 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.226020 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-rqsg6" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.226934 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.228169 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.237770 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.333931 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-796c85cffb-bqt8t"] Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.334184 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-796c85cffb-bqt8t" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerName="neutron-api" containerID="cri-o://7360b36476c2e0682a7873761f3f256f17cbf7a10797fd7003c2c5fa1c70362e" gracePeriod=30 Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.334926 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.335888 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-796c85cffb-bqt8t" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerName="neutron-httpd" containerID="cri-o://e0cb014851ced8c80e7c5be141d80f1fe266f43982b254effc4846669d5dbbdf" gracePeriod=30 Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.350767 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/da8872ad-7c19-451d-84be-6333126bfb63-openstack-config\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.350897 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/da8872ad-7c19-451d-84be-6333126bfb63-openstack-config-secret\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.350938 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhkch\" (UniqueName: \"kubernetes.io/projected/da8872ad-7c19-451d-84be-6333126bfb63-kube-api-access-hhkch\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.350966 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da8872ad-7c19-451d-84be-6333126bfb63-combined-ca-bundle\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.454921 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7z24\" (UniqueName: \"kubernetes.io/projected/486b072b-b195-440f-9d34-62a962bcbba6-kube-api-access-w7z24\") pod \"486b072b-b195-440f-9d34-62a962bcbba6\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455220 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-combined-ca-bundle\") pod \"486b072b-b195-440f-9d34-62a962bcbba6\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455255 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/486b072b-b195-440f-9d34-62a962bcbba6-etc-machine-id\") pod \"486b072b-b195-440f-9d34-62a962bcbba6\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455272 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-scripts\") pod \"486b072b-b195-440f-9d34-62a962bcbba6\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455515 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data\") pod \"486b072b-b195-440f-9d34-62a962bcbba6\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455537 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data-custom\") pod \"486b072b-b195-440f-9d34-62a962bcbba6\" (UID: \"486b072b-b195-440f-9d34-62a962bcbba6\") " Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455759 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/da8872ad-7c19-451d-84be-6333126bfb63-openstack-config\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455839 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/da8872ad-7c19-451d-84be-6333126bfb63-openstack-config-secret\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455885 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhkch\" (UniqueName: \"kubernetes.io/projected/da8872ad-7c19-451d-84be-6333126bfb63-kube-api-access-hhkch\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.455918 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da8872ad-7c19-451d-84be-6333126bfb63-combined-ca-bundle\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.458598 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/da8872ad-7c19-451d-84be-6333126bfb63-openstack-config\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.459705 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/486b072b-b195-440f-9d34-62a962bcbba6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "486b072b-b195-440f-9d34-62a962bcbba6" (UID: "486b072b-b195-440f-9d34-62a962bcbba6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.464780 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da8872ad-7c19-451d-84be-6333126bfb63-combined-ca-bundle\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.467737 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "486b072b-b195-440f-9d34-62a962bcbba6" (UID: "486b072b-b195-440f-9d34-62a962bcbba6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.468588 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/486b072b-b195-440f-9d34-62a962bcbba6-kube-api-access-w7z24" (OuterVolumeSpecName: "kube-api-access-w7z24") pod "486b072b-b195-440f-9d34-62a962bcbba6" (UID: "486b072b-b195-440f-9d34-62a962bcbba6"). InnerVolumeSpecName "kube-api-access-w7z24". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.468597 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-scripts" (OuterVolumeSpecName: "scripts") pod "486b072b-b195-440f-9d34-62a962bcbba6" (UID: "486b072b-b195-440f-9d34-62a962bcbba6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.472903 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/da8872ad-7c19-451d-84be-6333126bfb63-openstack-config-secret\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.476392 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhkch\" (UniqueName: \"kubernetes.io/projected/da8872ad-7c19-451d-84be-6333126bfb63-kube-api-access-hhkch\") pod \"openstackclient\" (UID: \"da8872ad-7c19-451d-84be-6333126bfb63\") " pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.557821 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7z24\" (UniqueName: \"kubernetes.io/projected/486b072b-b195-440f-9d34-62a962bcbba6-kube-api-access-w7z24\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.557852 4735 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/486b072b-b195-440f-9d34-62a962bcbba6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.557867 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.557879 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.574662 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "486b072b-b195-440f-9d34-62a962bcbba6" (UID: "486b072b-b195-440f-9d34-62a962bcbba6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.612264 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data" (OuterVolumeSpecName: "config-data") pod "486b072b-b195-440f-9d34-62a962bcbba6" (UID: "486b072b-b195-440f-9d34-62a962bcbba6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.641545 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.660180 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.660243 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/486b072b-b195-440f-9d34-62a962bcbba6-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:22 crc kubenswrapper[4735]: I1122 08:25:22.937329 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.152966 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"486b072b-b195-440f-9d34-62a962bcbba6","Type":"ContainerDied","Data":"9f3973eb1ab41ea061eba1944acb4acf61ad2297ae737bfe757f0f3d810818f4"} Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.153017 4735 scope.go:117] "RemoveContainer" containerID="43c2e6cad4d45b375688baf10460e72b6ddb138c67e3a13d62b310e381e3787c" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.153121 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.169939 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"da8872ad-7c19-451d-84be-6333126bfb63","Type":"ContainerStarted","Data":"32e2bab5af5b7c2178571f7e7113e3baf915e3b90f92ff52187d4e1975e8c286"} Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.173706 4735 generic.go:334] "Generic (PLEG): container finished" podID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerID="e0cb014851ced8c80e7c5be141d80f1fe266f43982b254effc4846669d5dbbdf" exitCode=0 Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.173734 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-796c85cffb-bqt8t" event={"ID":"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a","Type":"ContainerDied","Data":"e0cb014851ced8c80e7c5be141d80f1fe266f43982b254effc4846669d5dbbdf"} Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.197376 4735 scope.go:117] "RemoveContainer" containerID="8f7c6544be271f5b2aae181961fe1de597cb18b3210467d04623bbdfced54728" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.205532 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.227290 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.237706 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:23 crc kubenswrapper[4735]: E1122 08:25:23.238115 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="486b072b-b195-440f-9d34-62a962bcbba6" containerName="probe" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.238127 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="486b072b-b195-440f-9d34-62a962bcbba6" containerName="probe" Nov 22 08:25:23 crc kubenswrapper[4735]: E1122 08:25:23.238160 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="486b072b-b195-440f-9d34-62a962bcbba6" containerName="cinder-scheduler" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.238166 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="486b072b-b195-440f-9d34-62a962bcbba6" containerName="cinder-scheduler" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.238348 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="486b072b-b195-440f-9d34-62a962bcbba6" containerName="cinder-scheduler" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.238381 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="486b072b-b195-440f-9d34-62a962bcbba6" containerName="probe" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.239434 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.242872 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.249022 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.275705 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xwn9\" (UniqueName: \"kubernetes.io/projected/55ef656f-16ef-4187-9d57-01ab16e11c17-kube-api-access-5xwn9\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.275770 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-scripts\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.275792 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.275865 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.275884 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-config-data\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.275924 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55ef656f-16ef-4187-9d57-01ab16e11c17-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: E1122 08:25:23.276022 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod486b072b_b195_440f_9d34_62a962bcbba6.slice\": RecentStats: unable to find data in memory cache]" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.280157 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="486b072b-b195-440f-9d34-62a962bcbba6" path="/var/lib/kubelet/pods/486b072b-b195-440f-9d34-62a962bcbba6/volumes" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.378186 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xwn9\" (UniqueName: \"kubernetes.io/projected/55ef656f-16ef-4187-9d57-01ab16e11c17-kube-api-access-5xwn9\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.378243 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-scripts\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.378266 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.378367 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.378388 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-config-data\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.378426 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55ef656f-16ef-4187-9d57-01ab16e11c17-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.379973 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55ef656f-16ef-4187-9d57-01ab16e11c17-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.386095 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.388020 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.388054 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-scripts\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.388286 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ef656f-16ef-4187-9d57-01ab16e11c17-config-data\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.415241 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xwn9\" (UniqueName: \"kubernetes.io/projected/55ef656f-16ef-4187-9d57-01ab16e11c17-kube-api-access-5xwn9\") pod \"cinder-scheduler-0\" (UID: \"55ef656f-16ef-4187-9d57-01ab16e11c17\") " pod="openstack/cinder-scheduler-0" Nov 22 08:25:23 crc kubenswrapper[4735]: I1122 08:25:23.561136 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:25:24 crc kubenswrapper[4735]: I1122 08:25:24.081082 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:25:24 crc kubenswrapper[4735]: I1122 08:25:24.197477 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55ef656f-16ef-4187-9d57-01ab16e11c17","Type":"ContainerStarted","Data":"261ada6ad5351d22b40a9f5fc9a652e2f90e274788901f37a13f8c697ddb9f13"} Nov 22 08:25:25 crc kubenswrapper[4735]: I1122 08:25:25.260001 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55ef656f-16ef-4187-9d57-01ab16e11c17","Type":"ContainerStarted","Data":"699006d76a6fed860958f0a6c849e29591bf17c85d86fe7b5f69b982558f777c"} Nov 22 08:25:25 crc kubenswrapper[4735]: I1122 08:25:25.834830 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.321796 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55ef656f-16ef-4187-9d57-01ab16e11c17","Type":"ContainerStarted","Data":"376db2fd59b4fa52ed0a793f2fc80e19966e3dff30f7956b74c7e66b9e571ccf"} Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.342878 4735 generic.go:334] "Generic (PLEG): container finished" podID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerID="7360b36476c2e0682a7873761f3f256f17cbf7a10797fd7003c2c5fa1c70362e" exitCode=0 Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.342925 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-796c85cffb-bqt8t" event={"ID":"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a","Type":"ContainerDied","Data":"7360b36476c2e0682a7873761f3f256f17cbf7a10797fd7003c2c5fa1c70362e"} Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.370424 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.370400593 podStartE2EDuration="3.370400593s" podCreationTimestamp="2025-11-22 08:25:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:26.358136623 +0000 UTC m=+1347.962475228" watchObservedRunningTime="2025-11-22 08:25:26.370400593 +0000 UTC m=+1347.974739198" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.570032 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.738679 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-ovndb-tls-certs\") pod \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.738779 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-combined-ca-bundle\") pod \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.738832 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8r6d\" (UniqueName: \"kubernetes.io/projected/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-kube-api-access-s8r6d\") pod \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.738876 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-httpd-config\") pod \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.739047 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-config\") pod \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\" (UID: \"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a\") " Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.750319 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-kube-api-access-s8r6d" (OuterVolumeSpecName: "kube-api-access-s8r6d") pod "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" (UID: "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a"). InnerVolumeSpecName "kube-api-access-s8r6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.767653 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" (UID: "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.834810 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-config" (OuterVolumeSpecName: "config") pod "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" (UID: "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.842314 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.842343 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8r6d\" (UniqueName: \"kubernetes.io/projected/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-kube-api-access-s8r6d\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.842384 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.880778 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" (UID: "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.906388 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" (UID: "b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.944952 4735 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:26 crc kubenswrapper[4735]: I1122 08:25:26.945156 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.072560 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-d55fd4b4f-24s49"] Nov 22 08:25:27 crc kubenswrapper[4735]: E1122 08:25:27.073103 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerName="neutron-api" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.073126 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerName="neutron-api" Nov 22 08:25:27 crc kubenswrapper[4735]: E1122 08:25:27.073164 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerName="neutron-httpd" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.073173 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerName="neutron-httpd" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.076530 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerName="neutron-api" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.076585 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" containerName="neutron-httpd" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.078234 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.080412 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.080705 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.080777 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.092049 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-d55fd4b4f-24s49"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.251851 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-public-tls-certs\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.251927 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3fc589b-feac-438c-b008-8911417ccfa7-run-httpd\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.252176 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3fc589b-feac-438c-b008-8911417ccfa7-log-httpd\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.252436 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdmjg\" (UniqueName: \"kubernetes.io/projected/b3fc589b-feac-438c-b008-8911417ccfa7-kube-api-access-qdmjg\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.252597 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b3fc589b-feac-438c-b008-8911417ccfa7-etc-swift\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.252645 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-combined-ca-bundle\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.252820 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-internal-tls-certs\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.252862 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-config-data\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.354061 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pl47j"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.355059 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-config-data\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.355630 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.355732 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-public-tls-certs\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.355783 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3fc589b-feac-438c-b008-8911417ccfa7-run-httpd\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.355827 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3fc589b-feac-438c-b008-8911417ccfa7-log-httpd\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.355865 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdmjg\" (UniqueName: \"kubernetes.io/projected/b3fc589b-feac-438c-b008-8911417ccfa7-kube-api-access-qdmjg\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.355938 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b3fc589b-feac-438c-b008-8911417ccfa7-etc-swift\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.355955 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-combined-ca-bundle\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.356082 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-internal-tls-certs\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.359757 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3fc589b-feac-438c-b008-8911417ccfa7-run-httpd\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.360417 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-config-data\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.360424 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3fc589b-feac-438c-b008-8911417ccfa7-log-httpd\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.372514 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-public-tls-certs\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.374100 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-combined-ca-bundle\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.380279 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pl47j"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.383588 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b3fc589b-feac-438c-b008-8911417ccfa7-etc-swift\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.396531 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fc589b-feac-438c-b008-8911417ccfa7-internal-tls-certs\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.414216 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdmjg\" (UniqueName: \"kubernetes.io/projected/b3fc589b-feac-438c-b008-8911417ccfa7-kube-api-access-qdmjg\") pod \"swift-proxy-d55fd4b4f-24s49\" (UID: \"b3fc589b-feac-438c-b008-8911417ccfa7\") " pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.422763 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-796c85cffb-bqt8t" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.424014 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-796c85cffb-bqt8t" event={"ID":"b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a","Type":"ContainerDied","Data":"3f7ca1bb5c0d781ec8e26d6ec48cf79b89e2c7bc60dbbfbb57d274954c011f9a"} Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.424049 4735 scope.go:117] "RemoveContainer" containerID="e0cb014851ced8c80e7c5be141d80f1fe266f43982b254effc4846669d5dbbdf" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.458041 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/991c0a20-2e5f-441a-ac7f-a74d0a83814e-operator-scripts\") pod \"nova-api-db-create-pl47j\" (UID: \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\") " pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.458506 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f6bd\" (UniqueName: \"kubernetes.io/projected/991c0a20-2e5f-441a-ac7f-a74d0a83814e-kube-api-access-7f6bd\") pod \"nova-api-db-create-pl47j\" (UID: \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\") " pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.487686 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-89zq6"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.489228 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.507948 4735 scope.go:117] "RemoveContainer" containerID="7360b36476c2e0682a7873761f3f256f17cbf7a10797fd7003c2c5fa1c70362e" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.508316 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-796c85cffb-bqt8t"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.556477 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-d80d-account-create-lcmww"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.557855 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.561055 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/991c0a20-2e5f-441a-ac7f-a74d0a83814e-operator-scripts\") pod \"nova-api-db-create-pl47j\" (UID: \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\") " pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.561174 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f6bd\" (UniqueName: \"kubernetes.io/projected/991c0a20-2e5f-441a-ac7f-a74d0a83814e-kube-api-access-7f6bd\") pod \"nova-api-db-create-pl47j\" (UID: \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\") " pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.561261 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b854f12-c3ca-424a-8c7d-12d4d708174f-operator-scripts\") pod \"nova-cell0-db-create-89zq6\" (UID: \"8b854f12-c3ca-424a-8c7d-12d4d708174f\") " pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.561579 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8dzd\" (UniqueName: \"kubernetes.io/projected/8b854f12-c3ca-424a-8c7d-12d4d708174f-kube-api-access-t8dzd\") pod \"nova-cell0-db-create-89zq6\" (UID: \"8b854f12-c3ca-424a-8c7d-12d4d708174f\") " pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.563753 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/991c0a20-2e5f-441a-ac7f-a74d0a83814e-operator-scripts\") pod \"nova-api-db-create-pl47j\" (UID: \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\") " pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.564737 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.569518 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-796c85cffb-bqt8t"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.582390 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d80d-account-create-lcmww"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.588813 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f6bd\" (UniqueName: \"kubernetes.io/projected/991c0a20-2e5f-441a-ac7f-a74d0a83814e-kube-api-access-7f6bd\") pod \"nova-api-db-create-pl47j\" (UID: \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\") " pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.591617 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-89zq6"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.639001 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-npss9"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.640587 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.667835 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-npss9"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.700298 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.718659 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7bec6aa-ba3f-495e-bd32-52eb888657ac-operator-scripts\") pod \"nova-cell1-db-create-npss9\" (UID: \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\") " pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.718931 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-operator-scripts\") pod \"nova-api-d80d-account-create-lcmww\" (UID: \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\") " pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.719206 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-774rh\" (UniqueName: \"kubernetes.io/projected/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-kube-api-access-774rh\") pod \"nova-api-d80d-account-create-lcmww\" (UID: \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\") " pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.719490 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b854f12-c3ca-424a-8c7d-12d4d708174f-operator-scripts\") pod \"nova-cell0-db-create-89zq6\" (UID: \"8b854f12-c3ca-424a-8c7d-12d4d708174f\") " pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.719764 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rl2w\" (UniqueName: \"kubernetes.io/projected/d7bec6aa-ba3f-495e-bd32-52eb888657ac-kube-api-access-9rl2w\") pod \"nova-cell1-db-create-npss9\" (UID: \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\") " pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.720030 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8dzd\" (UniqueName: \"kubernetes.io/projected/8b854f12-c3ca-424a-8c7d-12d4d708174f-kube-api-access-t8dzd\") pod \"nova-cell0-db-create-89zq6\" (UID: \"8b854f12-c3ca-424a-8c7d-12d4d708174f\") " pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.721045 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b854f12-c3ca-424a-8c7d-12d4d708174f-operator-scripts\") pod \"nova-cell0-db-create-89zq6\" (UID: \"8b854f12-c3ca-424a-8c7d-12d4d708174f\") " pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.754884 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8dzd\" (UniqueName: \"kubernetes.io/projected/8b854f12-c3ca-424a-8c7d-12d4d708174f-kube-api-access-t8dzd\") pod \"nova-cell0-db-create-89zq6\" (UID: \"8b854f12-c3ca-424a-8c7d-12d4d708174f\") " pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.766536 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-16cf-account-create-7jcz2"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.771598 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-16cf-account-create-7jcz2"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.771927 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.786873 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.791198 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.822940 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rl2w\" (UniqueName: \"kubernetes.io/projected/d7bec6aa-ba3f-495e-bd32-52eb888657ac-kube-api-access-9rl2w\") pod \"nova-cell1-db-create-npss9\" (UID: \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\") " pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.823161 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7bec6aa-ba3f-495e-bd32-52eb888657ac-operator-scripts\") pod \"nova-cell1-db-create-npss9\" (UID: \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\") " pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.823199 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-operator-scripts\") pod \"nova-api-d80d-account-create-lcmww\" (UID: \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\") " pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.823263 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-774rh\" (UniqueName: \"kubernetes.io/projected/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-kube-api-access-774rh\") pod \"nova-api-d80d-account-create-lcmww\" (UID: \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\") " pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.824789 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7bec6aa-ba3f-495e-bd32-52eb888657ac-operator-scripts\") pod \"nova-cell1-db-create-npss9\" (UID: \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\") " pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.825137 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-operator-scripts\") pod \"nova-api-d80d-account-create-lcmww\" (UID: \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\") " pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.853672 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rl2w\" (UniqueName: \"kubernetes.io/projected/d7bec6aa-ba3f-495e-bd32-52eb888657ac-kube-api-access-9rl2w\") pod \"nova-cell1-db-create-npss9\" (UID: \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\") " pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.857550 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-774rh\" (UniqueName: \"kubernetes.io/projected/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-kube-api-access-774rh\") pod \"nova-api-d80d-account-create-lcmww\" (UID: \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\") " pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.858331 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.897025 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.899514 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-7129-account-create-8nxqz"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.901511 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.911227 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.917007 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7129-account-create-8nxqz"] Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.931790 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27d5c57d-8359-4b2f-96e5-d42f40634fd1-operator-scripts\") pod \"nova-cell0-16cf-account-create-7jcz2\" (UID: \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\") " pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.931974 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0eb1004-ded0-44d0-8341-f51eedd26af3-operator-scripts\") pod \"nova-cell1-7129-account-create-8nxqz\" (UID: \"a0eb1004-ded0-44d0-8341-f51eedd26af3\") " pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.932080 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzc4z\" (UniqueName: \"kubernetes.io/projected/a0eb1004-ded0-44d0-8341-f51eedd26af3-kube-api-access-vzc4z\") pod \"nova-cell1-7129-account-create-8nxqz\" (UID: \"a0eb1004-ded0-44d0-8341-f51eedd26af3\") " pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.932128 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x84h\" (UniqueName: \"kubernetes.io/projected/27d5c57d-8359-4b2f-96e5-d42f40634fd1-kube-api-access-4x84h\") pod \"nova-cell0-16cf-account-create-7jcz2\" (UID: \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\") " pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:27 crc kubenswrapper[4735]: I1122 08:25:27.983096 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.035282 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27d5c57d-8359-4b2f-96e5-d42f40634fd1-operator-scripts\") pod \"nova-cell0-16cf-account-create-7jcz2\" (UID: \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\") " pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.035427 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0eb1004-ded0-44d0-8341-f51eedd26af3-operator-scripts\") pod \"nova-cell1-7129-account-create-8nxqz\" (UID: \"a0eb1004-ded0-44d0-8341-f51eedd26af3\") " pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.035542 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzc4z\" (UniqueName: \"kubernetes.io/projected/a0eb1004-ded0-44d0-8341-f51eedd26af3-kube-api-access-vzc4z\") pod \"nova-cell1-7129-account-create-8nxqz\" (UID: \"a0eb1004-ded0-44d0-8341-f51eedd26af3\") " pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.035576 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x84h\" (UniqueName: \"kubernetes.io/projected/27d5c57d-8359-4b2f-96e5-d42f40634fd1-kube-api-access-4x84h\") pod \"nova-cell0-16cf-account-create-7jcz2\" (UID: \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\") " pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.036172 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27d5c57d-8359-4b2f-96e5-d42f40634fd1-operator-scripts\") pod \"nova-cell0-16cf-account-create-7jcz2\" (UID: \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\") " pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.036747 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0eb1004-ded0-44d0-8341-f51eedd26af3-operator-scripts\") pod \"nova-cell1-7129-account-create-8nxqz\" (UID: \"a0eb1004-ded0-44d0-8341-f51eedd26af3\") " pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.074086 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x84h\" (UniqueName: \"kubernetes.io/projected/27d5c57d-8359-4b2f-96e5-d42f40634fd1-kube-api-access-4x84h\") pod \"nova-cell0-16cf-account-create-7jcz2\" (UID: \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\") " pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.088084 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzc4z\" (UniqueName: \"kubernetes.io/projected/a0eb1004-ded0-44d0-8341-f51eedd26af3-kube-api-access-vzc4z\") pod \"nova-cell1-7129-account-create-8nxqz\" (UID: \"a0eb1004-ded0-44d0-8341-f51eedd26af3\") " pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.192334 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.203038 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.562061 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.692846 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-d55fd4b4f-24s49"] Nov 22 08:25:28 crc kubenswrapper[4735]: W1122 08:25:28.711038 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3fc589b_feac_438c_b008_8911417ccfa7.slice/crio-aa79632e6460fc98aa630033844079ac2633458cdd39104b2d64c46d1d539484 WatchSource:0}: Error finding container aa79632e6460fc98aa630033844079ac2633458cdd39104b2d64c46d1d539484: Status 404 returned error can't find the container with id aa79632e6460fc98aa630033844079ac2633458cdd39104b2d64c46d1d539484 Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.887055 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.887770 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="ceilometer-central-agent" containerID="cri-o://8ebb0c9292dcb1041f95e192c1b40e3b7a324dd2bb34acddf3dd4a05bf1feb2b" gracePeriod=30 Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.891137 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="proxy-httpd" containerID="cri-o://c8ea1500e428226bdd507961f97a1f2fc9ed4a47730c05233c5f59f53d35e882" gracePeriod=30 Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.891386 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="ceilometer-notification-agent" containerID="cri-o://cbcb10f4e1ead4c3a62dcdb5ecc619d242a4847559407a08570669f7b6c82553" gracePeriod=30 Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.891431 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="sg-core" containerID="cri-o://c149bb79a9740082c76c781ede88172aff86af44fbfbd02d97318e12f40634b1" gracePeriod=30 Nov 22 08:25:28 crc kubenswrapper[4735]: I1122 08:25:28.905730 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.202:3000/\": EOF" Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.056425 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pl47j"] Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.074050 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d80d-account-create-lcmww"] Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.081153 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-npss9"] Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.296982 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a" path="/var/lib/kubelet/pods/b478ffd5-1a7e-48ba-a5d3-5ba260ff7e7a/volumes" Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.459532 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-89zq6"] Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.516770 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pl47j" event={"ID":"991c0a20-2e5f-441a-ac7f-a74d0a83814e","Type":"ContainerStarted","Data":"b587d1927996fc10bf9e244fbbc159dd6b29178740c14203815ab50ff2e49517"} Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.545407 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-npss9" event={"ID":"d7bec6aa-ba3f-495e-bd32-52eb888657ac","Type":"ContainerStarted","Data":"5920bba4156992509fb16321310a110a8fd7c2671cbacb5e21bcb10e1e3c2315"} Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.556552 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-16cf-account-create-7jcz2"] Nov 22 08:25:29 crc kubenswrapper[4735]: W1122 08:25:29.575112 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0eb1004_ded0_44d0_8341_f51eedd26af3.slice/crio-6ec1276e8b5ff1ff9de6bf0021f8df33f15f23f2c0e38e3eb56fef04f1c2bbfa WatchSource:0}: Error finding container 6ec1276e8b5ff1ff9de6bf0021f8df33f15f23f2c0e38e3eb56fef04f1c2bbfa: Status 404 returned error can't find the container with id 6ec1276e8b5ff1ff9de6bf0021f8df33f15f23f2c0e38e3eb56fef04f1c2bbfa Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.589318 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7129-account-create-8nxqz"] Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.610141 4735 generic.go:334] "Generic (PLEG): container finished" podID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerID="c8ea1500e428226bdd507961f97a1f2fc9ed4a47730c05233c5f59f53d35e882" exitCode=0 Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.610177 4735 generic.go:334] "Generic (PLEG): container finished" podID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerID="c149bb79a9740082c76c781ede88172aff86af44fbfbd02d97318e12f40634b1" exitCode=2 Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.610220 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerDied","Data":"c8ea1500e428226bdd507961f97a1f2fc9ed4a47730c05233c5f59f53d35e882"} Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.610245 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerDied","Data":"c149bb79a9740082c76c781ede88172aff86af44fbfbd02d97318e12f40634b1"} Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.621806 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d80d-account-create-lcmww" event={"ID":"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88","Type":"ContainerStarted","Data":"4e2a87e2e56a740b9a4ebfd4fb37873eed603114f1c7eb215f466168fd5e7f80"} Nov 22 08:25:29 crc kubenswrapper[4735]: I1122 08:25:29.650172 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d55fd4b4f-24s49" event={"ID":"b3fc589b-feac-438c-b008-8911417ccfa7","Type":"ContainerStarted","Data":"aa79632e6460fc98aa630033844079ac2633458cdd39104b2d64c46d1d539484"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.674743 4735 generic.go:334] "Generic (PLEG): container finished" podID="345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88" containerID="9904d55dc8b1dee2eafe426aa9211ddcb4efb7cc5fe879cab7931d1ecb0659ed" exitCode=0 Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.674944 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d80d-account-create-lcmww" event={"ID":"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88","Type":"ContainerDied","Data":"9904d55dc8b1dee2eafe426aa9211ddcb4efb7cc5fe879cab7931d1ecb0659ed"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.681571 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d55fd4b4f-24s49" event={"ID":"b3fc589b-feac-438c-b008-8911417ccfa7","Type":"ContainerStarted","Data":"5049f9f5e9f2f4711cb291f2a704755dada83460b06e33dd23f30ae4b50cefdb"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.681623 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d55fd4b4f-24s49" event={"ID":"b3fc589b-feac-438c-b008-8911417ccfa7","Type":"ContainerStarted","Data":"d18616fd8009d220e9686e7ffb517a1a86a706b637eddc611905cffa7e9835aa"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.681668 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.681693 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.684439 4735 generic.go:334] "Generic (PLEG): container finished" podID="a0eb1004-ded0-44d0-8341-f51eedd26af3" containerID="fea186d92751736a4ba2f572e095a82e7d5c1ea21e90b189cf09b7ed91f41b8b" exitCode=0 Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.684515 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7129-account-create-8nxqz" event={"ID":"a0eb1004-ded0-44d0-8341-f51eedd26af3","Type":"ContainerDied","Data":"fea186d92751736a4ba2f572e095a82e7d5c1ea21e90b189cf09b7ed91f41b8b"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.684536 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7129-account-create-8nxqz" event={"ID":"a0eb1004-ded0-44d0-8341-f51eedd26af3","Type":"ContainerStarted","Data":"6ec1276e8b5ff1ff9de6bf0021f8df33f15f23f2c0e38e3eb56fef04f1c2bbfa"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.687388 4735 generic.go:334] "Generic (PLEG): container finished" podID="991c0a20-2e5f-441a-ac7f-a74d0a83814e" containerID="f94fa5d7222201202d8524076d62ab837ec52117dcc4bb104cd31b816c7731ca" exitCode=0 Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.687559 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pl47j" event={"ID":"991c0a20-2e5f-441a-ac7f-a74d0a83814e","Type":"ContainerDied","Data":"f94fa5d7222201202d8524076d62ab837ec52117dcc4bb104cd31b816c7731ca"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.695261 4735 generic.go:334] "Generic (PLEG): container finished" podID="27d5c57d-8359-4b2f-96e5-d42f40634fd1" containerID="bd0b109ae928074b6c650543bfc8b22b8acdabc52080177050d1e17b716fadfb" exitCode=0 Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.695363 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-16cf-account-create-7jcz2" event={"ID":"27d5c57d-8359-4b2f-96e5-d42f40634fd1","Type":"ContainerDied","Data":"bd0b109ae928074b6c650543bfc8b22b8acdabc52080177050d1e17b716fadfb"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.695395 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-16cf-account-create-7jcz2" event={"ID":"27d5c57d-8359-4b2f-96e5-d42f40634fd1","Type":"ContainerStarted","Data":"f1465fb0d8aac7986b5c6631f543ab6cbe6e672778899c07ab73cfc5481a5be1"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.703590 4735 generic.go:334] "Generic (PLEG): container finished" podID="d7bec6aa-ba3f-495e-bd32-52eb888657ac" containerID="e11ce79e7c417e089cbd02e1985b969c78b556e5bf67d6b391ff280846122cf8" exitCode=0 Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.703671 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-npss9" event={"ID":"d7bec6aa-ba3f-495e-bd32-52eb888657ac","Type":"ContainerDied","Data":"e11ce79e7c417e089cbd02e1985b969c78b556e5bf67d6b391ff280846122cf8"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.711222 4735 generic.go:334] "Generic (PLEG): container finished" podID="8b854f12-c3ca-424a-8c7d-12d4d708174f" containerID="29d6a4d2a95babc7f17622223035ec852d5c4ccd92902d75abc1b039c548c691" exitCode=0 Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.711377 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-89zq6" event={"ID":"8b854f12-c3ca-424a-8c7d-12d4d708174f","Type":"ContainerDied","Data":"29d6a4d2a95babc7f17622223035ec852d5c4ccd92902d75abc1b039c548c691"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.711437 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-89zq6" event={"ID":"8b854f12-c3ca-424a-8c7d-12d4d708174f","Type":"ContainerStarted","Data":"b5535090f3ee3dda711cccb8d7f12061536b1e5404d0999e1933310524e2415d"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.721174 4735 generic.go:334] "Generic (PLEG): container finished" podID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerID="8ebb0c9292dcb1041f95e192c1b40e3b7a324dd2bb34acddf3dd4a05bf1feb2b" exitCode=0 Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.721215 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerDied","Data":"8ebb0c9292dcb1041f95e192c1b40e3b7a324dd2bb34acddf3dd4a05bf1feb2b"} Nov 22 08:25:30 crc kubenswrapper[4735]: I1122 08:25:30.781591 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-d55fd4b4f-24s49" podStartSLOduration=3.781565438 podStartE2EDuration="3.781565438s" podCreationTimestamp="2025-11-22 08:25:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:30.737641386 +0000 UTC m=+1352.341980001" watchObservedRunningTime="2025-11-22 08:25:30.781565438 +0000 UTC m=+1352.385904043" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.786861 4735 generic.go:334] "Generic (PLEG): container finished" podID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerID="cbcb10f4e1ead4c3a62dcdb5ecc619d242a4847559407a08570669f7b6c82553" exitCode=0 Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.791374 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerDied","Data":"cbcb10f4e1ead4c3a62dcdb5ecc619d242a4847559407a08570669f7b6c82553"} Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.791420 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-555bb56d96-gvrkq"] Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.792903 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.801379 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.801414 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.801340 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-qpwzp" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.805278 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9wcd\" (UniqueName: \"kubernetes.io/projected/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-kube-api-access-h9wcd\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.805402 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-combined-ca-bundle\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.805530 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.805565 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data-custom\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.820481 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-555bb56d96-gvrkq"] Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.915630 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9wcd\" (UniqueName: \"kubernetes.io/projected/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-kube-api-access-h9wcd\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.915941 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-combined-ca-bundle\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.916019 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.916042 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data-custom\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.925939 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-lsxsl"] Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.927793 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.955498 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-combined-ca-bundle\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:32 crc kubenswrapper[4735]: I1122 08:25:32.985422 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data-custom\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.000451 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-lsxsl"] Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.009555 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.018985 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.019429 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.019502 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wv77\" (UniqueName: \"kubernetes.io/projected/0e32c687-6521-4512-8578-8c52fe8765d4-kube-api-access-4wv77\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.019538 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.019677 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-config\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.019880 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.023706 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9wcd\" (UniqueName: \"kubernetes.io/projected/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-kube-api-access-h9wcd\") pod \"heat-engine-555bb56d96-gvrkq\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.039639 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-99b96cbf9-4rfhb"] Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.042928 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.044636 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.070799 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-9d7f477d8-gpncg"] Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.072249 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.081940 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.102845 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-99b96cbf9-4rfhb"] Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.118013 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-9d7f477d8-gpncg"] Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121532 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmqxt\" (UniqueName: \"kubernetes.io/projected/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-kube-api-access-rmqxt\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121581 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-combined-ca-bundle\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121615 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data-custom\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121644 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121664 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zrnn\" (UniqueName: \"kubernetes.io/projected/b7bdb732-0307-475f-8224-9c5e3a3e66a6-kube-api-access-2zrnn\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121683 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121702 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data-custom\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121775 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121801 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121836 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121859 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wv77\" (UniqueName: \"kubernetes.io/projected/0e32c687-6521-4512-8578-8c52fe8765d4-kube-api-access-4wv77\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121878 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121931 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-config\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.121965 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-combined-ca-bundle\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.122933 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.123349 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.123444 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.124263 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-config\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.125053 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.131483 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.177476 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wv77\" (UniqueName: \"kubernetes.io/projected/0e32c687-6521-4512-8578-8c52fe8765d4-kube-api-access-4wv77\") pod \"dnsmasq-dns-7756b9d78c-lsxsl\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.225150 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmqxt\" (UniqueName: \"kubernetes.io/projected/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-kube-api-access-rmqxt\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.225198 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-combined-ca-bundle\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.225223 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data-custom\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.225258 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zrnn\" (UniqueName: \"kubernetes.io/projected/b7bdb732-0307-475f-8224-9c5e3a3e66a6-kube-api-access-2zrnn\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.225286 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.225305 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data-custom\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.225366 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.225723 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-combined-ca-bundle\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.232674 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-combined-ca-bundle\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.232835 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.233274 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data-custom\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.235024 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data-custom\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.242936 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-combined-ca-bundle\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.248447 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zrnn\" (UniqueName: \"kubernetes.io/projected/b7bdb732-0307-475f-8224-9c5e3a3e66a6-kube-api-access-2zrnn\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.249798 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data\") pod \"heat-api-99b96cbf9-4rfhb\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.255844 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmqxt\" (UniqueName: \"kubernetes.io/projected/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-kube-api-access-rmqxt\") pod \"heat-cfnapi-9d7f477d8-gpncg\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.342655 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.438117 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.518191 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:33 crc kubenswrapper[4735]: I1122 08:25:33.892481 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 22 08:25:37 crc kubenswrapper[4735]: I1122 08:25:37.713631 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:37 crc kubenswrapper[4735]: I1122 08:25:37.714360 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-d55fd4b4f-24s49" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.474882 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6796459cfd-kjmnc"] Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.476762 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.526360 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-8dbcbbcfc-4lcpv"] Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.529128 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.566912 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6796459cfd-kjmnc"] Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.587492 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-8dbcbbcfc-4lcpv"] Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.617904 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-cb7b4fb8c-qxxch"] Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.620302 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.631673 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-combined-ca-bundle\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.636555 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrqgs\" (UniqueName: \"kubernetes.io/projected/1db40e01-d3c8-4cc3-a571-f21049cd6616-kube-api-access-jrqgs\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.636843 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data-custom\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.637534 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.648600 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cb7b4fb8c-qxxch"] Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.657368 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.730267 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.738621 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8dzd\" (UniqueName: \"kubernetes.io/projected/8b854f12-c3ca-424a-8c7d-12d4d708174f-kube-api-access-t8dzd\") pod \"8b854f12-c3ca-424a-8c7d-12d4d708174f\" (UID: \"8b854f12-c3ca-424a-8c7d-12d4d708174f\") " Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.738712 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27d5c57d-8359-4b2f-96e5-d42f40634fd1-operator-scripts\") pod \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\" (UID: \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\") " Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.738782 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x84h\" (UniqueName: \"kubernetes.io/projected/27d5c57d-8359-4b2f-96e5-d42f40634fd1-kube-api-access-4x84h\") pod \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\" (UID: \"27d5c57d-8359-4b2f-96e5-d42f40634fd1\") " Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.738839 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b854f12-c3ca-424a-8c7d-12d4d708174f-operator-scripts\") pod \"8b854f12-c3ca-424a-8c7d-12d4d708174f\" (UID: \"8b854f12-c3ca-424a-8c7d-12d4d708174f\") " Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739024 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data-custom\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739053 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwvwf\" (UniqueName: \"kubernetes.io/projected/777b5609-9fd0-4781-a1db-dffa53b29764-kube-api-access-wwvwf\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739083 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739118 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data-custom\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739135 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data-custom\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739155 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-combined-ca-bundle\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739200 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739229 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrqgs\" (UniqueName: \"kubernetes.io/projected/1db40e01-d3c8-4cc3-a571-f21049cd6616-kube-api-access-jrqgs\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739247 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739277 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvwcl\" (UniqueName: \"kubernetes.io/projected/d71bec67-ffe7-4875-9145-d910676454d4-kube-api-access-mvwcl\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739309 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-combined-ca-bundle\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.739373 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-combined-ca-bundle\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.740366 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b854f12-c3ca-424a-8c7d-12d4d708174f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b854f12-c3ca-424a-8c7d-12d4d708174f" (UID: "8b854f12-c3ca-424a-8c7d-12d4d708174f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.749009 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27d5c57d-8359-4b2f-96e5-d42f40634fd1-kube-api-access-4x84h" (OuterVolumeSpecName: "kube-api-access-4x84h") pod "27d5c57d-8359-4b2f-96e5-d42f40634fd1" (UID: "27d5c57d-8359-4b2f-96e5-d42f40634fd1"). InnerVolumeSpecName "kube-api-access-4x84h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.750203 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data-custom\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.753607 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b854f12-c3ca-424a-8c7d-12d4d708174f-kube-api-access-t8dzd" (OuterVolumeSpecName: "kube-api-access-t8dzd") pod "8b854f12-c3ca-424a-8c7d-12d4d708174f" (UID: "8b854f12-c3ca-424a-8c7d-12d4d708174f"). InnerVolumeSpecName "kube-api-access-t8dzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.757806 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27d5c57d-8359-4b2f-96e5-d42f40634fd1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "27d5c57d-8359-4b2f-96e5-d42f40634fd1" (UID: "27d5c57d-8359-4b2f-96e5-d42f40634fd1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.764582 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrqgs\" (UniqueName: \"kubernetes.io/projected/1db40e01-d3c8-4cc3-a571-f21049cd6616-kube-api-access-jrqgs\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.774427 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.775119 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-combined-ca-bundle\") pod \"heat-engine-6796459cfd-kjmnc\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.815676 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.831765 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.846674 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data-custom\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.846715 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data-custom\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.846850 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.846909 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.846967 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvwcl\" (UniqueName: \"kubernetes.io/projected/d71bec67-ffe7-4875-9145-d910676454d4-kube-api-access-mvwcl\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.847007 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-combined-ca-bundle\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.847155 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-combined-ca-bundle\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.847315 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwvwf\" (UniqueName: \"kubernetes.io/projected/777b5609-9fd0-4781-a1db-dffa53b29764-kube-api-access-wwvwf\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.850315 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8dzd\" (UniqueName: \"kubernetes.io/projected/8b854f12-c3ca-424a-8c7d-12d4d708174f-kube-api-access-t8dzd\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.850337 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27d5c57d-8359-4b2f-96e5-d42f40634fd1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.850346 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x84h\" (UniqueName: \"kubernetes.io/projected/27d5c57d-8359-4b2f-96e5-d42f40634fd1-kube-api-access-4x84h\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.850356 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b854f12-c3ca-424a-8c7d-12d4d708174f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.857018 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data-custom\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.860726 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data-custom\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.867612 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-combined-ca-bundle\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.871879 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-combined-ca-bundle\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.872584 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvwcl\" (UniqueName: \"kubernetes.io/projected/d71bec67-ffe7-4875-9145-d910676454d4-kube-api-access-mvwcl\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.872772 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.881852 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwvwf\" (UniqueName: \"kubernetes.io/projected/777b5609-9fd0-4781-a1db-dffa53b29764-kube-api-access-wwvwf\") pod \"heat-api-8dbcbbcfc-4lcpv\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.888555 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data\") pod \"heat-cfnapi-cb7b4fb8c-qxxch\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.893211 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-npss9" event={"ID":"d7bec6aa-ba3f-495e-bd32-52eb888657ac","Type":"ContainerDied","Data":"5920bba4156992509fb16321310a110a8fd7c2671cbacb5e21bcb10e1e3c2315"} Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.893250 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5920bba4156992509fb16321310a110a8fd7c2671cbacb5e21bcb10e1e3c2315" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.897303 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-89zq6" event={"ID":"8b854f12-c3ca-424a-8c7d-12d4d708174f","Type":"ContainerDied","Data":"b5535090f3ee3dda711cccb8d7f12061536b1e5404d0999e1933310524e2415d"} Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.897517 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5535090f3ee3dda711cccb8d7f12061536b1e5404d0999e1933310524e2415d" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.897632 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-89zq6" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.902814 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.914909 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8d81c52-8a16-40a1-b2c2-1a476275b92c","Type":"ContainerDied","Data":"6a6cf54ba44c2d49464e7119d2b76890e1e3e0378f82ef10b37a29fc616779fa"} Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.914945 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a6cf54ba44c2d49464e7119d2b76890e1e3e0378f82ef10b37a29fc616779fa" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.916963 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d80d-account-create-lcmww" event={"ID":"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88","Type":"ContainerDied","Data":"4e2a87e2e56a740b9a4ebfd4fb37873eed603114f1c7eb215f466168fd5e7f80"} Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.916987 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e2a87e2e56a740b9a4ebfd4fb37873eed603114f1c7eb215f466168fd5e7f80" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.920163 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.923686 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7129-account-create-8nxqz" event={"ID":"a0eb1004-ded0-44d0-8341-f51eedd26af3","Type":"ContainerDied","Data":"6ec1276e8b5ff1ff9de6bf0021f8df33f15f23f2c0e38e3eb56fef04f1c2bbfa"} Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.924004 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ec1276e8b5ff1ff9de6bf0021f8df33f15f23f2c0e38e3eb56fef04f1c2bbfa" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.924154 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7129-account-create-8nxqz" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.932964 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.938425 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pl47j" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.938533 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pl47j" event={"ID":"991c0a20-2e5f-441a-ac7f-a74d0a83814e","Type":"ContainerDied","Data":"b587d1927996fc10bf9e244fbbc159dd6b29178740c14203815ab50ff2e49517"} Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.939598 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b587d1927996fc10bf9e244fbbc159dd6b29178740c14203815ab50ff2e49517" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.945779 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-16cf-account-create-7jcz2" event={"ID":"27d5c57d-8359-4b2f-96e5-d42f40634fd1","Type":"ContainerDied","Data":"f1465fb0d8aac7986b5c6631f543ab6cbe6e672778899c07ab73cfc5481a5be1"} Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.945815 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1465fb0d8aac7986b5c6631f543ab6cbe6e672778899c07ab73cfc5481a5be1" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.945862 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-16cf-account-create-7jcz2" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.955395 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/991c0a20-2e5f-441a-ac7f-a74d0a83814e-operator-scripts\") pod \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\" (UID: \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\") " Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.955837 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f6bd\" (UniqueName: \"kubernetes.io/projected/991c0a20-2e5f-441a-ac7f-a74d0a83814e-kube-api-access-7f6bd\") pod \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\" (UID: \"991c0a20-2e5f-441a-ac7f-a74d0a83814e\") " Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.956124 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0eb1004-ded0-44d0-8341-f51eedd26af3-operator-scripts\") pod \"a0eb1004-ded0-44d0-8341-f51eedd26af3\" (UID: \"a0eb1004-ded0-44d0-8341-f51eedd26af3\") " Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.956238 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzc4z\" (UniqueName: \"kubernetes.io/projected/a0eb1004-ded0-44d0-8341-f51eedd26af3-kube-api-access-vzc4z\") pod \"a0eb1004-ded0-44d0-8341-f51eedd26af3\" (UID: \"a0eb1004-ded0-44d0-8341-f51eedd26af3\") " Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.957175 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0eb1004-ded0-44d0-8341-f51eedd26af3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a0eb1004-ded0-44d0-8341-f51eedd26af3" (UID: "a0eb1004-ded0-44d0-8341-f51eedd26af3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.957528 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991c0a20-2e5f-441a-ac7f-a74d0a83814e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "991c0a20-2e5f-441a-ac7f-a74d0a83814e" (UID: "991c0a20-2e5f-441a-ac7f-a74d0a83814e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.959235 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0eb1004-ded0-44d0-8341-f51eedd26af3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.959331 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/991c0a20-2e5f-441a-ac7f-a74d0a83814e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.962015 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.970065 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/991c0a20-2e5f-441a-ac7f-a74d0a83814e-kube-api-access-7f6bd" (OuterVolumeSpecName: "kube-api-access-7f6bd") pod "991c0a20-2e5f-441a-ac7f-a74d0a83814e" (UID: "991c0a20-2e5f-441a-ac7f-a74d0a83814e"). InnerVolumeSpecName "kube-api-access-7f6bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.979637 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0eb1004-ded0-44d0-8341-f51eedd26af3-kube-api-access-vzc4z" (OuterVolumeSpecName: "kube-api-access-vzc4z") pod "a0eb1004-ded0-44d0-8341-f51eedd26af3" (UID: "a0eb1004-ded0-44d0-8341-f51eedd26af3"). InnerVolumeSpecName "kube-api-access-vzc4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:39 crc kubenswrapper[4735]: I1122 08:25:39.992838 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.012414 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.062507 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-774rh\" (UniqueName: \"kubernetes.io/projected/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-kube-api-access-774rh\") pod \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\" (UID: \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.062606 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-operator-scripts\") pod \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\" (UID: \"345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063406 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-sg-core-conf-yaml\") pod \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063469 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-config-data\") pod \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063505 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-log-httpd\") pod \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063537 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7bec6aa-ba3f-495e-bd32-52eb888657ac-operator-scripts\") pod \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\" (UID: \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063557 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fgpg\" (UniqueName: \"kubernetes.io/projected/a8d81c52-8a16-40a1-b2c2-1a476275b92c-kube-api-access-4fgpg\") pod \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063577 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rl2w\" (UniqueName: \"kubernetes.io/projected/d7bec6aa-ba3f-495e-bd32-52eb888657ac-kube-api-access-9rl2w\") pod \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\" (UID: \"d7bec6aa-ba3f-495e-bd32-52eb888657ac\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063585 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88" (UID: "345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063608 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-combined-ca-bundle\") pod \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063649 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-run-httpd\") pod \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.063685 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-scripts\") pod \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\" (UID: \"a8d81c52-8a16-40a1-b2c2-1a476275b92c\") " Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.064699 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a8d81c52-8a16-40a1-b2c2-1a476275b92c" (UID: "a8d81c52-8a16-40a1-b2c2-1a476275b92c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.064790 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7bec6aa-ba3f-495e-bd32-52eb888657ac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7bec6aa-ba3f-495e-bd32-52eb888657ac" (UID: "d7bec6aa-ba3f-495e-bd32-52eb888657ac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.066773 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a8d81c52-8a16-40a1-b2c2-1a476275b92c" (UID: "a8d81c52-8a16-40a1-b2c2-1a476275b92c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.067717 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8d81c52-8a16-40a1-b2c2-1a476275b92c-kube-api-access-4fgpg" (OuterVolumeSpecName: "kube-api-access-4fgpg") pod "a8d81c52-8a16-40a1-b2c2-1a476275b92c" (UID: "a8d81c52-8a16-40a1-b2c2-1a476275b92c"). InnerVolumeSpecName "kube-api-access-4fgpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.069730 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzc4z\" (UniqueName: \"kubernetes.io/projected/a0eb1004-ded0-44d0-8341-f51eedd26af3-kube-api-access-vzc4z\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.074952 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-scripts" (OuterVolumeSpecName: "scripts") pod "a8d81c52-8a16-40a1-b2c2-1a476275b92c" (UID: "a8d81c52-8a16-40a1-b2c2-1a476275b92c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.082017 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.082072 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.082087 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f6bd\" (UniqueName: \"kubernetes.io/projected/991c0a20-2e5f-441a-ac7f-a74d0a83814e-kube-api-access-7f6bd\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.082099 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8d81c52-8a16-40a1-b2c2-1a476275b92c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.082131 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fgpg\" (UniqueName: \"kubernetes.io/projected/a8d81c52-8a16-40a1-b2c2-1a476275b92c-kube-api-access-4fgpg\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.082143 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7bec6aa-ba3f-495e-bd32-52eb888657ac-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.082445 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-kube-api-access-774rh" (OuterVolumeSpecName: "kube-api-access-774rh") pod "345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88" (UID: "345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88"). InnerVolumeSpecName "kube-api-access-774rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.086190 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7bec6aa-ba3f-495e-bd32-52eb888657ac-kube-api-access-9rl2w" (OuterVolumeSpecName: "kube-api-access-9rl2w") pod "d7bec6aa-ba3f-495e-bd32-52eb888657ac" (UID: "d7bec6aa-ba3f-495e-bd32-52eb888657ac"). InnerVolumeSpecName "kube-api-access-9rl2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.123716 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a8d81c52-8a16-40a1-b2c2-1a476275b92c" (UID: "a8d81c52-8a16-40a1-b2c2-1a476275b92c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.185284 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.187120 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rl2w\" (UniqueName: \"kubernetes.io/projected/d7bec6aa-ba3f-495e-bd32-52eb888657ac-kube-api-access-9rl2w\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.187137 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.187148 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-774rh\" (UniqueName: \"kubernetes.io/projected/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88-kube-api-access-774rh\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.240838 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-555bb56d96-gvrkq"] Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.272972 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8d81c52-8a16-40a1-b2c2-1a476275b92c" (UID: "a8d81c52-8a16-40a1-b2c2-1a476275b92c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.295956 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.330753 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-99b96cbf9-4rfhb"] Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.372614 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-config-data" (OuterVolumeSpecName: "config-data") pod "a8d81c52-8a16-40a1-b2c2-1a476275b92c" (UID: "a8d81c52-8a16-40a1-b2c2-1a476275b92c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.398711 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d81c52-8a16-40a1-b2c2-1a476275b92c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:40 crc kubenswrapper[4735]: W1122 08:25:40.599289 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6a30c8d_c08e_47b0_bc92_99ba01d4f734.slice/crio-4854c6ad1890d8a039e10a2b08d14ba5d7bc618dfe04aefd89eff6e63d44bb74 WatchSource:0}: Error finding container 4854c6ad1890d8a039e10a2b08d14ba5d7bc618dfe04aefd89eff6e63d44bb74: Status 404 returned error can't find the container with id 4854c6ad1890d8a039e10a2b08d14ba5d7bc618dfe04aefd89eff6e63d44bb74 Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.604344 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-9d7f477d8-gpncg"] Nov 22 08:25:40 crc kubenswrapper[4735]: W1122 08:25:40.607577 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e32c687_6521_4512_8578_8c52fe8765d4.slice/crio-285d468c0081595a2e413cfad63fb1be48193e7f87b312dc0e6ed98cd03e9f40 WatchSource:0}: Error finding container 285d468c0081595a2e413cfad63fb1be48193e7f87b312dc0e6ed98cd03e9f40: Status 404 returned error can't find the container with id 285d468c0081595a2e413cfad63fb1be48193e7f87b312dc0e6ed98cd03e9f40 Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.625941 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-lsxsl"] Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.801971 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-8dbcbbcfc-4lcpv"] Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.876246 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6796459cfd-kjmnc"] Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.964063 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6796459cfd-kjmnc" event={"ID":"1db40e01-d3c8-4cc3-a571-f21049cd6616","Type":"ContainerStarted","Data":"7640a03a631afc4ef7a601f7a7103d87b499d0a0cc7ffad6ecfadb22991f6336"} Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.967041 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8dbcbbcfc-4lcpv" event={"ID":"777b5609-9fd0-4781-a1db-dffa53b29764","Type":"ContainerStarted","Data":"5834dd10e529241048350d119e362410a31b71af643114fbeb6f7b226bf20761"} Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.969236 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-555bb56d96-gvrkq" event={"ID":"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70","Type":"ContainerStarted","Data":"e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66"} Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.969287 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-555bb56d96-gvrkq" event={"ID":"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70","Type":"ContainerStarted","Data":"ceafd79e60f55d7695175b9e8ffd91933296c833c1b34fcfad7182b0b0ed4fc9"} Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.969306 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:40 crc kubenswrapper[4735]: I1122 08:25:40.974411 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-99b96cbf9-4rfhb" event={"ID":"b7bdb732-0307-475f-8224-9c5e3a3e66a6","Type":"ContainerStarted","Data":"afb1fac74cb4c069efe5152227942ccebf7a10dc86d153f133914a06857c789b"} Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.003209 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" event={"ID":"b6a30c8d-c08e-47b0-bc92-99ba01d4f734","Type":"ContainerStarted","Data":"4854c6ad1890d8a039e10a2b08d14ba5d7bc618dfe04aefd89eff6e63d44bb74"} Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.014838 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" event={"ID":"0e32c687-6521-4512-8578-8c52fe8765d4","Type":"ContainerStarted","Data":"546ffdf531969359975342136534ed420644c59deb0da3471b75bd6acfa81138"} Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.015041 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" event={"ID":"0e32c687-6521-4512-8578-8c52fe8765d4","Type":"ContainerStarted","Data":"285d468c0081595a2e413cfad63fb1be48193e7f87b312dc0e6ed98cd03e9f40"} Nov 22 08:25:41 crc kubenswrapper[4735]: W1122 08:25:41.025969 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd71bec67_ffe7_4875_9145_d910676454d4.slice/crio-a452c13fefa1d9adcb143455b663743a195e856ab949d6db67e56604d87f28a6 WatchSource:0}: Error finding container a452c13fefa1d9adcb143455b663743a195e856ab949d6db67e56604d87f28a6: Status 404 returned error can't find the container with id a452c13fefa1d9adcb143455b663743a195e856ab949d6db67e56604d87f28a6 Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.029781 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-npss9" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.030214 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"da8872ad-7c19-451d-84be-6333126bfb63","Type":"ContainerStarted","Data":"2919e65bc5748423e41cde47206f5f7e61c25f73a8d19f5c513580a41586a163"} Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.030564 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.031348 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d80d-account-create-lcmww" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.045335 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cb7b4fb8c-qxxch"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.063158 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-555bb56d96-gvrkq" podStartSLOduration=9.06313645 podStartE2EDuration="9.06313645s" podCreationTimestamp="2025-11-22 08:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:40.99450259 +0000 UTC m=+1362.598841195" watchObservedRunningTime="2025-11-22 08:25:41.06313645 +0000 UTC m=+1362.667475055" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.079072 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.601159957 podStartE2EDuration="19.079056953s" podCreationTimestamp="2025-11-22 08:25:22 +0000 UTC" firstStartedPulling="2025-11-22 08:25:22.965626579 +0000 UTC m=+1344.569965184" lastFinishedPulling="2025-11-22 08:25:39.443523575 +0000 UTC m=+1361.047862180" observedRunningTime="2025-11-22 08:25:41.055909468 +0000 UTC m=+1362.660248073" watchObservedRunningTime="2025-11-22 08:25:41.079056953 +0000 UTC m=+1362.683395558" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.119981 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.134623 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.150560 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151170 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="ceilometer-central-agent" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151194 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="ceilometer-central-agent" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151216 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="991c0a20-2e5f-441a-ac7f-a74d0a83814e" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151223 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="991c0a20-2e5f-441a-ac7f-a74d0a83814e" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151237 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="proxy-httpd" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151244 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="proxy-httpd" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151252 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="sg-core" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151260 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="sg-core" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151296 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7bec6aa-ba3f-495e-bd32-52eb888657ac" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151303 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7bec6aa-ba3f-495e-bd32-52eb888657ac" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151326 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0eb1004-ded0-44d0-8341-f51eedd26af3" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151335 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0eb1004-ded0-44d0-8341-f51eedd26af3" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151349 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b854f12-c3ca-424a-8c7d-12d4d708174f" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151356 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b854f12-c3ca-424a-8c7d-12d4d708174f" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151368 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d5c57d-8359-4b2f-96e5-d42f40634fd1" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151375 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d5c57d-8359-4b2f-96e5-d42f40634fd1" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151387 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151394 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: E1122 08:25:41.151403 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="ceilometer-notification-agent" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151409 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="ceilometer-notification-agent" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151682 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="ceilometer-notification-agent" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151706 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0eb1004-ded0-44d0-8341-f51eedd26af3" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151720 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="991c0a20-2e5f-441a-ac7f-a74d0a83814e" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151734 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="proxy-httpd" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151749 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="ceilometer-central-agent" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151764 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7bec6aa-ba3f-495e-bd32-52eb888657ac" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151777 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="sg-core" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151794 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151807 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d5c57d-8359-4b2f-96e5-d42f40634fd1" containerName="mariadb-account-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.151822 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b854f12-c3ca-424a-8c7d-12d4d708174f" containerName="mariadb-database-create" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.159003 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.163838 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.163942 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.166945 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.217007 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-log-httpd\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.217202 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.217241 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wdjn\" (UniqueName: \"kubernetes.io/projected/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-kube-api-access-4wdjn\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.217280 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-config-data\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.217348 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-scripts\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.217377 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-run-httpd\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.217449 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.292279 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" path="/var/lib/kubelet/pods/a8d81c52-8a16-40a1-b2c2-1a476275b92c/volumes" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.319501 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-log-httpd\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.319608 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.319633 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wdjn\" (UniqueName: \"kubernetes.io/projected/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-kube-api-access-4wdjn\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.319664 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-config-data\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.319713 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-scripts\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.319739 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-run-httpd\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.319795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.320649 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-run-httpd\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.320926 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-log-httpd\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.324239 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.328687 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.328880 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-config-data\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.329398 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-scripts\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.341161 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wdjn\" (UniqueName: \"kubernetes.io/projected/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-kube-api-access-4wdjn\") pod \"ceilometer-0\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.491721 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.882524 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-99b96cbf9-4rfhb"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.909715 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-9d7f477d8-gpncg"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.925820 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-b64b95447-gmvn2"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.927499 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.935024 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.935832 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.940534 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-b64b95447-gmvn2"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.959090 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7d54dc7bc4-cfqcw"] Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.960634 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.966863 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.967076 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Nov 22 08:25:41 crc kubenswrapper[4735]: I1122 08:25:41.986274 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7d54dc7bc4-cfqcw"] Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039244 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039312 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-public-tls-certs\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039340 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-internal-tls-certs\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039450 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-combined-ca-bundle\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039576 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nq88\" (UniqueName: \"kubernetes.io/projected/5d438585-a36f-4e86-ab10-df80ac1f8340-kube-api-access-4nq88\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039626 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-combined-ca-bundle\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039654 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data-custom\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039673 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039722 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-public-tls-certs\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039740 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data-custom\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039761 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6jfh\" (UniqueName: \"kubernetes.io/projected/06d273c9-b1e2-46bc-a6e3-26b26d24e966-kube-api-access-r6jfh\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.039811 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-internal-tls-certs\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.051973 4735 generic.go:334] "Generic (PLEG): container finished" podID="0e32c687-6521-4512-8578-8c52fe8765d4" containerID="546ffdf531969359975342136534ed420644c59deb0da3471b75bd6acfa81138" exitCode=0 Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.052041 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" event={"ID":"0e32c687-6521-4512-8578-8c52fe8765d4","Type":"ContainerDied","Data":"546ffdf531969359975342136534ed420644c59deb0da3471b75bd6acfa81138"} Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.052065 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" event={"ID":"0e32c687-6521-4512-8578-8c52fe8765d4","Type":"ContainerStarted","Data":"01da5ec42d9c40d97f8ff2d6a29f264b5b3089d8820d4baebaba235ac8adc066"} Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.053243 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.059975 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" event={"ID":"d71bec67-ffe7-4875-9145-d910676454d4","Type":"ContainerStarted","Data":"a452c13fefa1d9adcb143455b663743a195e856ab949d6db67e56604d87f28a6"} Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.063217 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6796459cfd-kjmnc" event={"ID":"1db40e01-d3c8-4cc3-a571-f21049cd6616","Type":"ContainerStarted","Data":"0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d"} Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.063247 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.074414 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" podStartSLOduration=10.074397875 podStartE2EDuration="10.074397875s" podCreationTimestamp="2025-11-22 08:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:42.073744317 +0000 UTC m=+1363.678082942" watchObservedRunningTime="2025-11-22 08:25:42.074397875 +0000 UTC m=+1363.678736470" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.099602 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6796459cfd-kjmnc" podStartSLOduration=3.099581175 podStartE2EDuration="3.099581175s" podCreationTimestamp="2025-11-22 08:25:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:42.089793553 +0000 UTC m=+1363.694132158" watchObservedRunningTime="2025-11-22 08:25:42.099581175 +0000 UTC m=+1363.703919780" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.142413 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-internal-tls-certs\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.142519 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.142642 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-public-tls-certs\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.143496 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-internal-tls-certs\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.143756 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-combined-ca-bundle\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.144948 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nq88\" (UniqueName: \"kubernetes.io/projected/5d438585-a36f-4e86-ab10-df80ac1f8340-kube-api-access-4nq88\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.145142 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-combined-ca-bundle\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.145423 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.145609 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data-custom\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.146152 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-public-tls-certs\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.146212 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data-custom\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.146249 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6jfh\" (UniqueName: \"kubernetes.io/projected/06d273c9-b1e2-46bc-a6e3-26b26d24e966-kube-api-access-r6jfh\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.154423 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-combined-ca-bundle\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.154535 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-internal-tls-certs\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.156433 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.157883 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-public-tls-certs\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.164808 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-combined-ca-bundle\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.165000 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data-custom\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.166295 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.167379 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data-custom\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.168165 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-internal-tls-certs\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.168987 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6jfh\" (UniqueName: \"kubernetes.io/projected/06d273c9-b1e2-46bc-a6e3-26b26d24e966-kube-api-access-r6jfh\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.171594 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-public-tls-certs\") pod \"heat-cfnapi-7d54dc7bc4-cfqcw\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.176547 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nq88\" (UniqueName: \"kubernetes.io/projected/5d438585-a36f-4e86-ab10-df80ac1f8340-kube-api-access-4nq88\") pod \"heat-api-b64b95447-gmvn2\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.259621 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:42 crc kubenswrapper[4735]: I1122 08:25:42.287808 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.517826 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sjwjn"] Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.520349 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.521775 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.526927 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.527216 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tvwv2" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.531579 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sjwjn"] Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.576533 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-config-data\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.576839 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-scripts\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.576974 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.577011 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8mmv\" (UniqueName: \"kubernetes.io/projected/6c3a99da-447d-4586-b5d8-2214d6df2cbd-kube-api-access-s8mmv\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.678596 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-scripts\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.678692 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.678721 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8mmv\" (UniqueName: \"kubernetes.io/projected/6c3a99da-447d-4586-b5d8-2214d6df2cbd-kube-api-access-s8mmv\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.678904 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-config-data\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.687609 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-config-data\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.689631 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.695629 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-scripts\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.709154 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8mmv\" (UniqueName: \"kubernetes.io/projected/6c3a99da-447d-4586-b5d8-2214d6df2cbd-kube-api-access-s8mmv\") pod \"nova-cell0-conductor-db-sync-sjwjn\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:43 crc kubenswrapper[4735]: I1122 08:25:43.849138 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:25:45 crc kubenswrapper[4735]: I1122 08:25:45.452434 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:45 crc kubenswrapper[4735]: I1122 08:25:45.668599 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7d54dc7bc4-cfqcw"] Nov 22 08:25:45 crc kubenswrapper[4735]: I1122 08:25:45.935166 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sjwjn"] Nov 22 08:25:45 crc kubenswrapper[4735]: W1122 08:25:45.945375 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d438585_a36f_4e86_ab10_df80ac1f8340.slice/crio-81ce07b78d21ab6d845110ef21de502a0577062d88ca4a37e61feeee912ac6e9 WatchSource:0}: Error finding container 81ce07b78d21ab6d845110ef21de502a0577062d88ca4a37e61feeee912ac6e9: Status 404 returned error can't find the container with id 81ce07b78d21ab6d845110ef21de502a0577062d88ca4a37e61feeee912ac6e9 Nov 22 08:25:45 crc kubenswrapper[4735]: I1122 08:25:45.969156 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-b64b95447-gmvn2"] Nov 22 08:25:46 crc kubenswrapper[4735]: I1122 08:25:46.153680 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" event={"ID":"06d273c9-b1e2-46bc-a6e3-26b26d24e966","Type":"ContainerStarted","Data":"219edf1a7a052034cee00fc7c095c0ddf6bbdc98efda0649760cb027b91ac394"} Nov 22 08:25:46 crc kubenswrapper[4735]: I1122 08:25:46.157940 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerStarted","Data":"38bec181b83170774e1b788e3301cb28929f4caa0d3c3c596d80b340dfdc40ed"} Nov 22 08:25:46 crc kubenswrapper[4735]: I1122 08:25:46.164140 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-b64b95447-gmvn2" event={"ID":"5d438585-a36f-4e86-ab10-df80ac1f8340","Type":"ContainerStarted","Data":"81ce07b78d21ab6d845110ef21de502a0577062d88ca4a37e61feeee912ac6e9"} Nov 22 08:25:46 crc kubenswrapper[4735]: I1122 08:25:46.167531 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" event={"ID":"6c3a99da-447d-4586-b5d8-2214d6df2cbd","Type":"ContainerStarted","Data":"65f2a834c3ff28c4fad2f09341d73811206972b81890195f166cdee947d4b61b"} Nov 22 08:25:46 crc kubenswrapper[4735]: I1122 08:25:46.174906 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-99b96cbf9-4rfhb" event={"ID":"b7bdb732-0307-475f-8224-9c5e3a3e66a6","Type":"ContainerStarted","Data":"f469819db7cac1e2222c8ce9694ba77c7f27cafc3ab708ff5b94553b1a7f114a"} Nov 22 08:25:46 crc kubenswrapper[4735]: I1122 08:25:46.175045 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-99b96cbf9-4rfhb" podUID="b7bdb732-0307-475f-8224-9c5e3a3e66a6" containerName="heat-api" containerID="cri-o://f469819db7cac1e2222c8ce9694ba77c7f27cafc3ab708ff5b94553b1a7f114a" gracePeriod=60 Nov 22 08:25:46 crc kubenswrapper[4735]: I1122 08:25:46.175311 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:46 crc kubenswrapper[4735]: I1122 08:25:46.199598 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-99b96cbf9-4rfhb" podStartSLOduration=9.458089583 podStartE2EDuration="14.199580194s" podCreationTimestamp="2025-11-22 08:25:32 +0000 UTC" firstStartedPulling="2025-11-22 08:25:40.323636083 +0000 UTC m=+1361.927974698" lastFinishedPulling="2025-11-22 08:25:45.065126704 +0000 UTC m=+1366.669465309" observedRunningTime="2025-11-22 08:25:46.197199139 +0000 UTC m=+1367.801537744" watchObservedRunningTime="2025-11-22 08:25:46.199580194 +0000 UTC m=+1367.803918799" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.471888 4735 generic.go:334] "Generic (PLEG): container finished" podID="d71bec67-ffe7-4875-9145-d910676454d4" containerID="0f475e5d55909e3d7f4389a29b60460ae83e916130a485dd0ff0aac62619e71a" exitCode=1 Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.472340 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" event={"ID":"d71bec67-ffe7-4875-9145-d910676454d4","Type":"ContainerDied","Data":"0f475e5d55909e3d7f4389a29b60460ae83e916130a485dd0ff0aac62619e71a"} Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.473068 4735 scope.go:117] "RemoveContainer" containerID="0f475e5d55909e3d7f4389a29b60460ae83e916130a485dd0ff0aac62619e71a" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.477095 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8dbcbbcfc-4lcpv" event={"ID":"777b5609-9fd0-4781-a1db-dffa53b29764","Type":"ContainerStarted","Data":"4160762f34e816a072886c11aee48578cc68408a4f9f038e873f9be337c765d1"} Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.477188 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.503476 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" event={"ID":"06d273c9-b1e2-46bc-a6e3-26b26d24e966","Type":"ContainerStarted","Data":"c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2"} Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.503838 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.516618 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerStarted","Data":"b469a9e6ff5a08892757f334bfbcbf8482dac49298107cbb11f5735e82507ae5"} Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.527133 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-b64b95447-gmvn2" event={"ID":"5d438585-a36f-4e86-ab10-df80ac1f8340","Type":"ContainerStarted","Data":"6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873"} Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.528093 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.533651 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" event={"ID":"b6a30c8d-c08e-47b0-bc92-99ba01d4f734","Type":"ContainerStarted","Data":"5137ef5adae744e18e159448ae9b962e0f5a181afc8b653d4f017de4f48d4073"} Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.534209 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" podUID="b6a30c8d-c08e-47b0-bc92-99ba01d4f734" containerName="heat-cfnapi" containerID="cri-o://5137ef5adae744e18e159448ae9b962e0f5a181afc8b653d4f017de4f48d4073" gracePeriod=60 Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.534415 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.534558 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-8dbcbbcfc-4lcpv" podStartSLOduration=4.18162754 podStartE2EDuration="8.534527578s" podCreationTimestamp="2025-11-22 08:25:39 +0000 UTC" firstStartedPulling="2025-11-22 08:25:40.812510819 +0000 UTC m=+1362.416849424" lastFinishedPulling="2025-11-22 08:25:45.165410857 +0000 UTC m=+1366.769749462" observedRunningTime="2025-11-22 08:25:47.517119352 +0000 UTC m=+1369.121457957" watchObservedRunningTime="2025-11-22 08:25:47.534527578 +0000 UTC m=+1369.138866183" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.555150 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" podStartSLOduration=6.555126501 podStartE2EDuration="6.555126501s" podCreationTimestamp="2025-11-22 08:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:47.531921456 +0000 UTC m=+1369.136260061" watchObservedRunningTime="2025-11-22 08:25:47.555126501 +0000 UTC m=+1369.159465106" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.576379 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-b64b95447-gmvn2" podStartSLOduration=6.5763575119999995 podStartE2EDuration="6.576357512s" podCreationTimestamp="2025-11-22 08:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:25:47.546841584 +0000 UTC m=+1369.151180209" watchObservedRunningTime="2025-11-22 08:25:47.576357512 +0000 UTC m=+1369.180696117" Nov 22 08:25:47 crc kubenswrapper[4735]: I1122 08:25:47.587452 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" podStartSLOduration=10.457449886 podStartE2EDuration="15.587433445s" podCreationTimestamp="2025-11-22 08:25:32 +0000 UTC" firstStartedPulling="2025-11-22 08:25:40.603790084 +0000 UTC m=+1362.208128679" lastFinishedPulling="2025-11-22 08:25:45.733773633 +0000 UTC m=+1367.338112238" observedRunningTime="2025-11-22 08:25:47.567917241 +0000 UTC m=+1369.172255846" watchObservedRunningTime="2025-11-22 08:25:47.587433445 +0000 UTC m=+1369.191772050" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.440606 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.509358 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5f5xm"] Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.509605 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" podUID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" containerName="dnsmasq-dns" containerID="cri-o://9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a" gracePeriod=10 Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.613973 4735 generic.go:334] "Generic (PLEG): container finished" podID="777b5609-9fd0-4781-a1db-dffa53b29764" containerID="4160762f34e816a072886c11aee48578cc68408a4f9f038e873f9be337c765d1" exitCode=1 Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.614180 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8dbcbbcfc-4lcpv" event={"ID":"777b5609-9fd0-4781-a1db-dffa53b29764","Type":"ContainerDied","Data":"4160762f34e816a072886c11aee48578cc68408a4f9f038e873f9be337c765d1"} Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.615430 4735 scope.go:117] "RemoveContainer" containerID="4160762f34e816a072886c11aee48578cc68408a4f9f038e873f9be337c765d1" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.650811 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerStarted","Data":"da36175fbcaf8cd07a106ded62d8e880079a7dc7476028ddda46700dba7b76e4"} Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.683753 4735 generic.go:334] "Generic (PLEG): container finished" podID="b6a30c8d-c08e-47b0-bc92-99ba01d4f734" containerID="5137ef5adae744e18e159448ae9b962e0f5a181afc8b653d4f017de4f48d4073" exitCode=0 Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.683853 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" event={"ID":"b6a30c8d-c08e-47b0-bc92-99ba01d4f734","Type":"ContainerDied","Data":"5137ef5adae744e18e159448ae9b962e0f5a181afc8b653d4f017de4f48d4073"} Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.700039 4735 generic.go:334] "Generic (PLEG): container finished" podID="d71bec67-ffe7-4875-9145-d910676454d4" containerID="0be5dcf6030de540f6b60ca6a6eae8aa2d11ee24ffb54899aa910dec0a93147d" exitCode=1 Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.702011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" event={"ID":"d71bec67-ffe7-4875-9145-d910676454d4","Type":"ContainerDied","Data":"0be5dcf6030de540f6b60ca6a6eae8aa2d11ee24ffb54899aa910dec0a93147d"} Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.702087 4735 scope.go:117] "RemoveContainer" containerID="0f475e5d55909e3d7f4389a29b60460ae83e916130a485dd0ff0aac62619e71a" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.714553 4735 scope.go:117] "RemoveContainer" containerID="0be5dcf6030de540f6b60ca6a6eae8aa2d11ee24ffb54899aa910dec0a93147d" Nov 22 08:25:48 crc kubenswrapper[4735]: E1122 08:25:48.714992 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-cb7b4fb8c-qxxch_openstack(d71bec67-ffe7-4875-9145-d910676454d4)\"" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" podUID="d71bec67-ffe7-4875-9145-d910676454d4" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.818421 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.892703 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-combined-ca-bundle\") pod \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.892796 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data-custom\") pod \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.892838 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmqxt\" (UniqueName: \"kubernetes.io/projected/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-kube-api-access-rmqxt\") pod \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.892899 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data\") pod \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\" (UID: \"b6a30c8d-c08e-47b0-bc92-99ba01d4f734\") " Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.909984 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b6a30c8d-c08e-47b0-bc92-99ba01d4f734" (UID: "b6a30c8d-c08e-47b0-bc92-99ba01d4f734"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.913443 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-kube-api-access-rmqxt" (OuterVolumeSpecName: "kube-api-access-rmqxt") pod "b6a30c8d-c08e-47b0-bc92-99ba01d4f734" (UID: "b6a30c8d-c08e-47b0-bc92-99ba01d4f734"). InnerVolumeSpecName "kube-api-access-rmqxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.955338 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6a30c8d-c08e-47b0-bc92-99ba01d4f734" (UID: "b6a30c8d-c08e-47b0-bc92-99ba01d4f734"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.994988 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.995051 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:48 crc kubenswrapper[4735]: I1122 08:25:48.995064 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmqxt\" (UniqueName: \"kubernetes.io/projected/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-kube-api-access-rmqxt\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.005045 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data" (OuterVolumeSpecName: "config-data") pod "b6a30c8d-c08e-47b0-bc92-99ba01d4f734" (UID: "b6a30c8d-c08e-47b0-bc92-99ba01d4f734"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.096979 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6a30c8d-c08e-47b0-bc92-99ba01d4f734-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.132711 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.200428 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-svc\") pod \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.200792 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r5m9\" (UniqueName: \"kubernetes.io/projected/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-kube-api-access-7r5m9\") pod \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.201001 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-sb\") pod \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.201112 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-config\") pod \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.201256 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-nb\") pod \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.201517 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-swift-storage-0\") pod \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\" (UID: \"f73d0ea1-05cf-413c-b5f8-2bef40df95b8\") " Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.222991 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-kube-api-access-7r5m9" (OuterVolumeSpecName: "kube-api-access-7r5m9") pod "f73d0ea1-05cf-413c-b5f8-2bef40df95b8" (UID: "f73d0ea1-05cf-413c-b5f8-2bef40df95b8"). InnerVolumeSpecName "kube-api-access-7r5m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.275229 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f73d0ea1-05cf-413c-b5f8-2bef40df95b8" (UID: "f73d0ea1-05cf-413c-b5f8-2bef40df95b8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.303263 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r5m9\" (UniqueName: \"kubernetes.io/projected/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-kube-api-access-7r5m9\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.303297 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.328284 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-config" (OuterVolumeSpecName: "config") pod "f73d0ea1-05cf-413c-b5f8-2bef40df95b8" (UID: "f73d0ea1-05cf-413c-b5f8-2bef40df95b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.365003 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f73d0ea1-05cf-413c-b5f8-2bef40df95b8" (UID: "f73d0ea1-05cf-413c-b5f8-2bef40df95b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.372820 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f73d0ea1-05cf-413c-b5f8-2bef40df95b8" (UID: "f73d0ea1-05cf-413c-b5f8-2bef40df95b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.405818 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.405854 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.405864 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.466245 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f73d0ea1-05cf-413c-b5f8-2bef40df95b8" (UID: "f73d0ea1-05cf-413c-b5f8-2bef40df95b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.510145 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d0ea1-05cf-413c-b5f8-2bef40df95b8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.721744 4735 scope.go:117] "RemoveContainer" containerID="0be5dcf6030de540f6b60ca6a6eae8aa2d11ee24ffb54899aa910dec0a93147d" Nov 22 08:25:49 crc kubenswrapper[4735]: E1122 08:25:49.722061 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-cb7b4fb8c-qxxch_openstack(d71bec67-ffe7-4875-9145-d910676454d4)\"" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" podUID="d71bec67-ffe7-4875-9145-d910676454d4" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.729932 4735 generic.go:334] "Generic (PLEG): container finished" podID="777b5609-9fd0-4781-a1db-dffa53b29764" containerID="dc72afa3b0d5137866f26b8083bff20be8136f57e644c0af2cf4891b359623c6" exitCode=1 Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.729988 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8dbcbbcfc-4lcpv" event={"ID":"777b5609-9fd0-4781-a1db-dffa53b29764","Type":"ContainerDied","Data":"dc72afa3b0d5137866f26b8083bff20be8136f57e644c0af2cf4891b359623c6"} Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.730078 4735 scope.go:117] "RemoveContainer" containerID="4160762f34e816a072886c11aee48578cc68408a4f9f038e873f9be337c765d1" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.730877 4735 scope.go:117] "RemoveContainer" containerID="dc72afa3b0d5137866f26b8083bff20be8136f57e644c0af2cf4891b359623c6" Nov 22 08:25:49 crc kubenswrapper[4735]: E1122 08:25:49.731229 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8dbcbbcfc-4lcpv_openstack(777b5609-9fd0-4781-a1db-dffa53b29764)\"" pod="openstack/heat-api-8dbcbbcfc-4lcpv" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.751034 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerStarted","Data":"78dd47744ff42aee374d5ad49ae8eb2c62c176f49f2b64186fa01bebbeccd25b"} Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.759159 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" event={"ID":"b6a30c8d-c08e-47b0-bc92-99ba01d4f734","Type":"ContainerDied","Data":"4854c6ad1890d8a039e10a2b08d14ba5d7bc618dfe04aefd89eff6e63d44bb74"} Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.759255 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9d7f477d8-gpncg" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.772147 4735 generic.go:334] "Generic (PLEG): container finished" podID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" containerID="9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a" exitCode=0 Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.772336 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" event={"ID":"f73d0ea1-05cf-413c-b5f8-2bef40df95b8","Type":"ContainerDied","Data":"9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a"} Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.772566 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" event={"ID":"f73d0ea1-05cf-413c-b5f8-2bef40df95b8","Type":"ContainerDied","Data":"cf36dacc03afd6d1451b24e76e6be33a4764831dcde1c654f45d8c5ab8a18093"} Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.772436 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5f5xm" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.822526 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-9d7f477d8-gpncg"] Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.836386 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-9d7f477d8-gpncg"] Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.836687 4735 scope.go:117] "RemoveContainer" containerID="5137ef5adae744e18e159448ae9b962e0f5a181afc8b653d4f017de4f48d4073" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.850316 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5f5xm"] Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.862128 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5f5xm"] Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.883656 4735 scope.go:117] "RemoveContainer" containerID="9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.957703 4735 scope.go:117] "RemoveContainer" containerID="c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.994581 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:49 crc kubenswrapper[4735]: I1122 08:25:49.996717 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:50 crc kubenswrapper[4735]: I1122 08:25:50.010623 4735 scope.go:117] "RemoveContainer" containerID="9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a" Nov 22 08:25:50 crc kubenswrapper[4735]: I1122 08:25:50.012703 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:50 crc kubenswrapper[4735]: I1122 08:25:50.012753 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:50 crc kubenswrapper[4735]: E1122 08:25:50.034894 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a\": container with ID starting with 9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a not found: ID does not exist" containerID="9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a" Nov 22 08:25:50 crc kubenswrapper[4735]: I1122 08:25:50.034932 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a"} err="failed to get container status \"9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a\": rpc error: code = NotFound desc = could not find container \"9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a\": container with ID starting with 9fd12db8c6ad7844e7d6d6c9e04125dfb250014fa44367d21051b7e43671857a not found: ID does not exist" Nov 22 08:25:50 crc kubenswrapper[4735]: I1122 08:25:50.034956 4735 scope.go:117] "RemoveContainer" containerID="c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150" Nov 22 08:25:50 crc kubenswrapper[4735]: E1122 08:25:50.035671 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150\": container with ID starting with c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150 not found: ID does not exist" containerID="c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150" Nov 22 08:25:50 crc kubenswrapper[4735]: I1122 08:25:50.035694 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150"} err="failed to get container status \"c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150\": rpc error: code = NotFound desc = could not find container \"c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150\": container with ID starting with c78dc08a8cfe971d800ad5ecb51eec8e004afd65b97ea308d332ba4162b70150 not found: ID does not exist" Nov 22 08:25:50 crc kubenswrapper[4735]: I1122 08:25:50.794515 4735 scope.go:117] "RemoveContainer" containerID="dc72afa3b0d5137866f26b8083bff20be8136f57e644c0af2cf4891b359623c6" Nov 22 08:25:50 crc kubenswrapper[4735]: E1122 08:25:50.795626 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8dbcbbcfc-4lcpv_openstack(777b5609-9fd0-4781-a1db-dffa53b29764)\"" pod="openstack/heat-api-8dbcbbcfc-4lcpv" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" Nov 22 08:25:50 crc kubenswrapper[4735]: I1122 08:25:50.798036 4735 scope.go:117] "RemoveContainer" containerID="0be5dcf6030de540f6b60ca6a6eae8aa2d11ee24ffb54899aa910dec0a93147d" Nov 22 08:25:50 crc kubenswrapper[4735]: E1122 08:25:50.798280 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-cb7b4fb8c-qxxch_openstack(d71bec67-ffe7-4875-9145-d910676454d4)\"" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" podUID="d71bec67-ffe7-4875-9145-d910676454d4" Nov 22 08:25:51 crc kubenswrapper[4735]: I1122 08:25:51.279494 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6a30c8d-c08e-47b0-bc92-99ba01d4f734" path="/var/lib/kubelet/pods/b6a30c8d-c08e-47b0-bc92-99ba01d4f734/volumes" Nov 22 08:25:51 crc kubenswrapper[4735]: I1122 08:25:51.291052 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" path="/var/lib/kubelet/pods/f73d0ea1-05cf-413c-b5f8-2bef40df95b8/volumes" Nov 22 08:25:51 crc kubenswrapper[4735]: I1122 08:25:51.813625 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerStarted","Data":"14b4b3e3dd41f2b911525d3192e343b6a2a67001193bb5212da8909f19a034cd"} Nov 22 08:25:51 crc kubenswrapper[4735]: I1122 08:25:51.814421 4735 scope.go:117] "RemoveContainer" containerID="dc72afa3b0d5137866f26b8083bff20be8136f57e644c0af2cf4891b359623c6" Nov 22 08:25:51 crc kubenswrapper[4735]: I1122 08:25:51.814673 4735 scope.go:117] "RemoveContainer" containerID="0be5dcf6030de540f6b60ca6a6eae8aa2d11ee24ffb54899aa910dec0a93147d" Nov 22 08:25:51 crc kubenswrapper[4735]: E1122 08:25:51.814727 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8dbcbbcfc-4lcpv_openstack(777b5609-9fd0-4781-a1db-dffa53b29764)\"" pod="openstack/heat-api-8dbcbbcfc-4lcpv" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" Nov 22 08:25:51 crc kubenswrapper[4735]: E1122 08:25:51.814893 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-cb7b4fb8c-qxxch_openstack(d71bec67-ffe7-4875-9145-d910676454d4)\"" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" podUID="d71bec67-ffe7-4875-9145-d910676454d4" Nov 22 08:25:51 crc kubenswrapper[4735]: I1122 08:25:51.839836 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.596194671 podStartE2EDuration="10.839813146s" podCreationTimestamp="2025-11-22 08:25:41 +0000 UTC" firstStartedPulling="2025-11-22 08:25:45.731154451 +0000 UTC m=+1367.335493056" lastFinishedPulling="2025-11-22 08:25:50.974772936 +0000 UTC m=+1372.579111531" observedRunningTime="2025-11-22 08:25:51.835768925 +0000 UTC m=+1373.440107540" watchObservedRunningTime="2025-11-22 08:25:51.839813146 +0000 UTC m=+1373.444151751" Nov 22 08:25:52 crc kubenswrapper[4735]: I1122 08:25:52.793553 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:25:52 crc kubenswrapper[4735]: I1122 08:25:52.824484 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:25:53 crc kubenswrapper[4735]: I1122 08:25:53.178240 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:25:53 crc kubenswrapper[4735]: I1122 08:25:53.835512 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="proxy-httpd" containerID="cri-o://14b4b3e3dd41f2b911525d3192e343b6a2a67001193bb5212da8909f19a034cd" gracePeriod=30 Nov 22 08:25:53 crc kubenswrapper[4735]: I1122 08:25:53.835545 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="sg-core" containerID="cri-o://78dd47744ff42aee374d5ad49ae8eb2c62c176f49f2b64186fa01bebbeccd25b" gracePeriod=30 Nov 22 08:25:53 crc kubenswrapper[4735]: I1122 08:25:53.835595 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="ceilometer-notification-agent" containerID="cri-o://da36175fbcaf8cd07a106ded62d8e880079a7dc7476028ddda46700dba7b76e4" gracePeriod=30 Nov 22 08:25:53 crc kubenswrapper[4735]: I1122 08:25:53.835449 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="ceilometer-central-agent" containerID="cri-o://b469a9e6ff5a08892757f334bfbcbf8482dac49298107cbb11f5735e82507ae5" gracePeriod=30 Nov 22 08:25:54 crc kubenswrapper[4735]: I1122 08:25:54.853446 4735 generic.go:334] "Generic (PLEG): container finished" podID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerID="14b4b3e3dd41f2b911525d3192e343b6a2a67001193bb5212da8909f19a034cd" exitCode=0 Nov 22 08:25:54 crc kubenswrapper[4735]: I1122 08:25:54.853748 4735 generic.go:334] "Generic (PLEG): container finished" podID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerID="78dd47744ff42aee374d5ad49ae8eb2c62c176f49f2b64186fa01bebbeccd25b" exitCode=2 Nov 22 08:25:54 crc kubenswrapper[4735]: I1122 08:25:54.853762 4735 generic.go:334] "Generic (PLEG): container finished" podID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerID="da36175fbcaf8cd07a106ded62d8e880079a7dc7476028ddda46700dba7b76e4" exitCode=0 Nov 22 08:25:54 crc kubenswrapper[4735]: I1122 08:25:54.853480 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerDied","Data":"14b4b3e3dd41f2b911525d3192e343b6a2a67001193bb5212da8909f19a034cd"} Nov 22 08:25:54 crc kubenswrapper[4735]: I1122 08:25:54.853796 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerDied","Data":"78dd47744ff42aee374d5ad49ae8eb2c62c176f49f2b64186fa01bebbeccd25b"} Nov 22 08:25:54 crc kubenswrapper[4735]: I1122 08:25:54.853811 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerDied","Data":"da36175fbcaf8cd07a106ded62d8e880079a7dc7476028ddda46700dba7b76e4"} Nov 22 08:25:55 crc kubenswrapper[4735]: I1122 08:25:55.644823 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:25:55 crc kubenswrapper[4735]: I1122 08:25:55.650313 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:25:55 crc kubenswrapper[4735]: I1122 08:25:55.745391 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-cb7b4fb8c-qxxch"] Nov 22 08:25:55 crc kubenswrapper[4735]: I1122 08:25:55.777693 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:25:55 crc kubenswrapper[4735]: I1122 08:25:55.852570 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-8dbcbbcfc-4lcpv"] Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.294738 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.304679 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.432642 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-combined-ca-bundle\") pod \"d71bec67-ffe7-4875-9145-d910676454d4\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.432727 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data\") pod \"777b5609-9fd0-4781-a1db-dffa53b29764\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.432778 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data-custom\") pod \"777b5609-9fd0-4781-a1db-dffa53b29764\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.432901 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvwcl\" (UniqueName: \"kubernetes.io/projected/d71bec67-ffe7-4875-9145-d910676454d4-kube-api-access-mvwcl\") pod \"d71bec67-ffe7-4875-9145-d910676454d4\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.432948 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data-custom\") pod \"d71bec67-ffe7-4875-9145-d910676454d4\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.433070 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-combined-ca-bundle\") pod \"777b5609-9fd0-4781-a1db-dffa53b29764\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.433108 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwvwf\" (UniqueName: \"kubernetes.io/projected/777b5609-9fd0-4781-a1db-dffa53b29764-kube-api-access-wwvwf\") pod \"777b5609-9fd0-4781-a1db-dffa53b29764\" (UID: \"777b5609-9fd0-4781-a1db-dffa53b29764\") " Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.433144 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data\") pod \"d71bec67-ffe7-4875-9145-d910676454d4\" (UID: \"d71bec67-ffe7-4875-9145-d910676454d4\") " Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.441616 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d71bec67-ffe7-4875-9145-d910676454d4" (UID: "d71bec67-ffe7-4875-9145-d910676454d4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.441939 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "777b5609-9fd0-4781-a1db-dffa53b29764" (UID: "777b5609-9fd0-4781-a1db-dffa53b29764"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.464711 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/777b5609-9fd0-4781-a1db-dffa53b29764-kube-api-access-wwvwf" (OuterVolumeSpecName: "kube-api-access-wwvwf") pod "777b5609-9fd0-4781-a1db-dffa53b29764" (UID: "777b5609-9fd0-4781-a1db-dffa53b29764"). InnerVolumeSpecName "kube-api-access-wwvwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.464832 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d71bec67-ffe7-4875-9145-d910676454d4-kube-api-access-mvwcl" (OuterVolumeSpecName: "kube-api-access-mvwcl") pod "d71bec67-ffe7-4875-9145-d910676454d4" (UID: "d71bec67-ffe7-4875-9145-d910676454d4"). InnerVolumeSpecName "kube-api-access-mvwcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.525604 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d71bec67-ffe7-4875-9145-d910676454d4" (UID: "d71bec67-ffe7-4875-9145-d910676454d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.535943 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.535973 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvwcl\" (UniqueName: \"kubernetes.io/projected/d71bec67-ffe7-4875-9145-d910676454d4-kube-api-access-mvwcl\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.535983 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.535991 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwvwf\" (UniqueName: \"kubernetes.io/projected/777b5609-9fd0-4781-a1db-dffa53b29764-kube-api-access-wwvwf\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.535999 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.549692 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data" (OuterVolumeSpecName: "config-data") pod "777b5609-9fd0-4781-a1db-dffa53b29764" (UID: "777b5609-9fd0-4781-a1db-dffa53b29764"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.567572 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data" (OuterVolumeSpecName: "config-data") pod "d71bec67-ffe7-4875-9145-d910676454d4" (UID: "d71bec67-ffe7-4875-9145-d910676454d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.581321 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "777b5609-9fd0-4781-a1db-dffa53b29764" (UID: "777b5609-9fd0-4781-a1db-dffa53b29764"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.638228 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.638256 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71bec67-ffe7-4875-9145-d910676454d4-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.638266 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777b5609-9fd0-4781-a1db-dffa53b29764-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.929604 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" event={"ID":"d71bec67-ffe7-4875-9145-d910676454d4","Type":"ContainerDied","Data":"a452c13fefa1d9adcb143455b663743a195e856ab949d6db67e56604d87f28a6"} Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.929662 4735 scope.go:117] "RemoveContainer" containerID="0be5dcf6030de540f6b60ca6a6eae8aa2d11ee24ffb54899aa910dec0a93147d" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.929666 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cb7b4fb8c-qxxch" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.934192 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8dbcbbcfc-4lcpv" event={"ID":"777b5609-9fd0-4781-a1db-dffa53b29764","Type":"ContainerDied","Data":"5834dd10e529241048350d119e362410a31b71af643114fbeb6f7b226bf20761"} Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.934272 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8dbcbbcfc-4lcpv" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.963892 4735 scope.go:117] "RemoveContainer" containerID="dc72afa3b0d5137866f26b8083bff20be8136f57e644c0af2cf4891b359623c6" Nov 22 08:25:58 crc kubenswrapper[4735]: I1122 08:25:58.985570 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-cb7b4fb8c-qxxch"] Nov 22 08:25:59 crc kubenswrapper[4735]: I1122 08:25:59.001367 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-cb7b4fb8c-qxxch"] Nov 22 08:25:59 crc kubenswrapper[4735]: I1122 08:25:59.023275 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-8dbcbbcfc-4lcpv"] Nov 22 08:25:59 crc kubenswrapper[4735]: I1122 08:25:59.040935 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-8dbcbbcfc-4lcpv"] Nov 22 08:25:59 crc kubenswrapper[4735]: I1122 08:25:59.294710 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" path="/var/lib/kubelet/pods/777b5609-9fd0-4781-a1db-dffa53b29764/volumes" Nov 22 08:25:59 crc kubenswrapper[4735]: I1122 08:25:59.295862 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d71bec67-ffe7-4875-9145-d910676454d4" path="/var/lib/kubelet/pods/d71bec67-ffe7-4875-9145-d910676454d4/volumes" Nov 22 08:25:59 crc kubenswrapper[4735]: I1122 08:25:59.969830 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" event={"ID":"6c3a99da-447d-4586-b5d8-2214d6df2cbd","Type":"ContainerStarted","Data":"859217d0a75b1a510d7d855ad2228d96c4008c74831dac4111be6e84d7bbd55b"} Nov 22 08:25:59 crc kubenswrapper[4735]: I1122 08:25:59.988544 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" podStartSLOduration=4.043650012 podStartE2EDuration="16.988529651s" podCreationTimestamp="2025-11-22 08:25:43 +0000 UTC" firstStartedPulling="2025-11-22 08:25:45.979015012 +0000 UTC m=+1367.583353617" lastFinishedPulling="2025-11-22 08:25:58.923894661 +0000 UTC m=+1380.528233256" observedRunningTime="2025-11-22 08:25:59.986377732 +0000 UTC m=+1381.590716347" watchObservedRunningTime="2025-11-22 08:25:59.988529651 +0000 UTC m=+1381.592868256" Nov 22 08:26:00 crc kubenswrapper[4735]: I1122 08:26:00.005286 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:26:00 crc kubenswrapper[4735]: I1122 08:26:00.070548 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-555bb56d96-gvrkq"] Nov 22 08:26:00 crc kubenswrapper[4735]: I1122 08:26:00.070977 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-555bb56d96-gvrkq" podUID="ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" containerName="heat-engine" containerID="cri-o://e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66" gracePeriod=60 Nov 22 08:26:03 crc kubenswrapper[4735]: E1122 08:26:03.135308 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:26:03 crc kubenswrapper[4735]: E1122 08:26:03.143225 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:26:03 crc kubenswrapper[4735]: E1122 08:26:03.147608 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:26:03 crc kubenswrapper[4735]: E1122 08:26:03.147701 4735 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-555bb56d96-gvrkq" podUID="ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" containerName="heat-engine" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.024630 4735 generic.go:334] "Generic (PLEG): container finished" podID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerID="b469a9e6ff5a08892757f334bfbcbf8482dac49298107cbb11f5735e82507ae5" exitCode=0 Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.024705 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerDied","Data":"b469a9e6ff5a08892757f334bfbcbf8482dac49298107cbb11f5735e82507ae5"} Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.025148 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5","Type":"ContainerDied","Data":"38bec181b83170774e1b788e3301cb28929f4caa0d3c3c596d80b340dfdc40ed"} Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.025168 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38bec181b83170774e1b788e3301cb28929f4caa0d3c3c596d80b340dfdc40ed" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.095008 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.215873 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-config-data\") pod \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.216298 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-sg-core-conf-yaml\") pod \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.216426 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-scripts\") pod \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.216563 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-run-httpd\") pod \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.216703 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-combined-ca-bundle\") pod \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.216953 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wdjn\" (UniqueName: \"kubernetes.io/projected/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-kube-api-access-4wdjn\") pod \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.217117 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" (UID: "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.217249 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-log-httpd\") pod \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\" (UID: \"72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5\") " Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.217874 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" (UID: "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.218154 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.218245 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.223906 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-scripts" (OuterVolumeSpecName: "scripts") pod "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" (UID: "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.229068 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-kube-api-access-4wdjn" (OuterVolumeSpecName: "kube-api-access-4wdjn") pod "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" (UID: "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5"). InnerVolumeSpecName "kube-api-access-4wdjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.273550 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" (UID: "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.320271 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.320299 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.320308 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wdjn\" (UniqueName: \"kubernetes.io/projected/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-kube-api-access-4wdjn\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.359628 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" (UID: "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.390649 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-config-data" (OuterVolumeSpecName: "config-data") pod "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" (UID: "72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.424187 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:05 crc kubenswrapper[4735]: I1122 08:26:05.424252 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.033991 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.074388 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.088401 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.101150 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.101811 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="ceilometer-central-agent" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.101842 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="ceilometer-central-agent" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.101865 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6a30c8d-c08e-47b0-bc92-99ba01d4f734" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.101875 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6a30c8d-c08e-47b0-bc92-99ba01d4f734" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.101895 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="proxy-httpd" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.101903 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="proxy-httpd" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.101920 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="sg-core" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.101928 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="sg-core" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.101945 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="ceilometer-notification-agent" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.101954 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="ceilometer-notification-agent" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.101972 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" containerName="heat-api" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.101981 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" containerName="heat-api" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.101994 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" containerName="dnsmasq-dns" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102003 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" containerName="dnsmasq-dns" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.102015 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d71bec67-ffe7-4875-9145-d910676454d4" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102024 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d71bec67-ffe7-4875-9145-d910676454d4" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.102046 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d71bec67-ffe7-4875-9145-d910676454d4" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102054 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d71bec67-ffe7-4875-9145-d910676454d4" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.102078 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" containerName="init" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102085 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" containerName="init" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102385 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" containerName="heat-api" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102418 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f73d0ea1-05cf-413c-b5f8-2bef40df95b8" containerName="dnsmasq-dns" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102442 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="ceilometer-notification-agent" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102520 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" containerName="heat-api" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102537 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d71bec67-ffe7-4875-9145-d910676454d4" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102557 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="sg-core" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102575 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6a30c8d-c08e-47b0-bc92-99ba01d4f734" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102584 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d71bec67-ffe7-4875-9145-d910676454d4" containerName="heat-cfnapi" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102598 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="ceilometer-central-agent" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102618 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" containerName="proxy-httpd" Nov 22 08:26:06 crc kubenswrapper[4735]: E1122 08:26:06.102895 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" containerName="heat-api" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.102906 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="777b5609-9fd0-4781-a1db-dffa53b29764" containerName="heat-api" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.105682 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.112622 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.122563 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.127618 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.245979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.246126 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-config-data\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.246154 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.246265 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.246346 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgfwt\" (UniqueName: \"kubernetes.io/projected/05987ef0-923f-428a-a7ea-968d8980c4b0-kube-api-access-lgfwt\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.246386 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-scripts\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.246521 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.348580 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.348685 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-config-data\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.348726 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.348807 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.348873 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgfwt\" (UniqueName: \"kubernetes.io/projected/05987ef0-923f-428a-a7ea-968d8980c4b0-kube-api-access-lgfwt\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.348930 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-scripts\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.349711 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.349734 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.350039 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.354262 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.355283 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-config-data\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.355406 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-scripts\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.356314 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.367389 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgfwt\" (UniqueName: \"kubernetes.io/projected/05987ef0-923f-428a-a7ea-968d8980c4b0-kube-api-access-lgfwt\") pod \"ceilometer-0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.435089 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:06 crc kubenswrapper[4735]: I1122 08:26:06.967045 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:07 crc kubenswrapper[4735]: I1122 08:26:07.047195 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerStarted","Data":"e908052ec606f1b79673f0d32d69115b33bb69594b21d31bba4f7a5c41ccef6f"} Nov 22 08:26:07 crc kubenswrapper[4735]: I1122 08:26:07.278675 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5" path="/var/lib/kubelet/pods/72817ac3-93ba-4d2b-a2bd-5c92ae07c8e5/volumes" Nov 22 08:26:07 crc kubenswrapper[4735]: I1122 08:26:07.281327 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a8d81c52-8a16-40a1-b2c2-1a476275b92c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.202:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:26:08 crc kubenswrapper[4735]: I1122 08:26:08.061491 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerStarted","Data":"df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e"} Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.086825 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerStarted","Data":"f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a"} Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.690894 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.825373 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data\") pod \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.825513 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9wcd\" (UniqueName: \"kubernetes.io/projected/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-kube-api-access-h9wcd\") pod \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.825618 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-combined-ca-bundle\") pod \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.826746 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data-custom\") pod \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\" (UID: \"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70\") " Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.830893 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" (UID: "ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.844242 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-kube-api-access-h9wcd" (OuterVolumeSpecName: "kube-api-access-h9wcd") pod "ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" (UID: "ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70"). InnerVolumeSpecName "kube-api-access-h9wcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.881738 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" (UID: "ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.899557 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data" (OuterVolumeSpecName: "config-data") pod "ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" (UID: "ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.929281 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9wcd\" (UniqueName: \"kubernetes.io/projected/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-kube-api-access-h9wcd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.929326 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.929339 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:09 crc kubenswrapper[4735]: I1122 08:26:09.929354 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.122043 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerStarted","Data":"203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe"} Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.129152 4735 generic.go:334] "Generic (PLEG): container finished" podID="ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" containerID="e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66" exitCode=0 Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.129212 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-555bb56d96-gvrkq" event={"ID":"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70","Type":"ContainerDied","Data":"e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66"} Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.129246 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-555bb56d96-gvrkq" event={"ID":"ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70","Type":"ContainerDied","Data":"ceafd79e60f55d7695175b9e8ffd91933296c833c1b34fcfad7182b0b0ed4fc9"} Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.129268 4735 scope.go:117] "RemoveContainer" containerID="e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66" Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.129561 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-555bb56d96-gvrkq" Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.211099 4735 scope.go:117] "RemoveContainer" containerID="e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66" Nov 22 08:26:10 crc kubenswrapper[4735]: E1122 08:26:10.212065 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66\": container with ID starting with e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66 not found: ID does not exist" containerID="e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66" Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.212108 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66"} err="failed to get container status \"e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66\": rpc error: code = NotFound desc = could not find container \"e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66\": container with ID starting with e9683180c08481d004d82d8b84f0f5093d725e7571c83b1c48977b7b5dafbe66 not found: ID does not exist" Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.217654 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-555bb56d96-gvrkq"] Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.244410 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-555bb56d96-gvrkq"] Nov 22 08:26:10 crc kubenswrapper[4735]: I1122 08:26:10.256982 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:11 crc kubenswrapper[4735]: I1122 08:26:11.314965 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" path="/var/lib/kubelet/pods/ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70/volumes" Nov 22 08:26:11 crc kubenswrapper[4735]: I1122 08:26:11.328282 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:26:11 crc kubenswrapper[4735]: I1122 08:26:11.328995 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-httpd" containerID="cri-o://ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d" gracePeriod=30 Nov 22 08:26:11 crc kubenswrapper[4735]: I1122 08:26:11.328574 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-log" containerID="cri-o://b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a" gracePeriod=30 Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.152117 4735 generic.go:334] "Generic (PLEG): container finished" podID="701597f5-e695-485b-a08c-cec90dad760f" containerID="b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a" exitCode=143 Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.152226 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"701597f5-e695-485b-a08c-cec90dad760f","Type":"ContainerDied","Data":"b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a"} Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.155593 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerStarted","Data":"21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57"} Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.155770 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="sg-core" containerID="cri-o://203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe" gracePeriod=30 Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.155780 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="proxy-httpd" containerID="cri-o://21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57" gracePeriod=30 Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.155794 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="ceilometer-notification-agent" containerID="cri-o://f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a" gracePeriod=30 Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.155840 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.156085 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="ceilometer-central-agent" containerID="cri-o://df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e" gracePeriod=30 Nov 22 08:26:12 crc kubenswrapper[4735]: I1122 08:26:12.180620 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.4478294099999998 podStartE2EDuration="6.18060518s" podCreationTimestamp="2025-11-22 08:26:06 +0000 UTC" firstStartedPulling="2025-11-22 08:26:06.971124619 +0000 UTC m=+1388.575463224" lastFinishedPulling="2025-11-22 08:26:10.703900389 +0000 UTC m=+1392.308238994" observedRunningTime="2025-11-22 08:26:12.176666841 +0000 UTC m=+1393.781005446" watchObservedRunningTime="2025-11-22 08:26:12.18060518 +0000 UTC m=+1393.784943785" Nov 22 08:26:13 crc kubenswrapper[4735]: I1122 08:26:13.171415 4735 generic.go:334] "Generic (PLEG): container finished" podID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerID="21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57" exitCode=0 Nov 22 08:26:13 crc kubenswrapper[4735]: I1122 08:26:13.171805 4735 generic.go:334] "Generic (PLEG): container finished" podID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerID="203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe" exitCode=2 Nov 22 08:26:13 crc kubenswrapper[4735]: I1122 08:26:13.171818 4735 generic.go:334] "Generic (PLEG): container finished" podID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerID="f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a" exitCode=0 Nov 22 08:26:13 crc kubenswrapper[4735]: I1122 08:26:13.171502 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerDied","Data":"21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57"} Nov 22 08:26:13 crc kubenswrapper[4735]: I1122 08:26:13.171865 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerDied","Data":"203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe"} Nov 22 08:26:13 crc kubenswrapper[4735]: I1122 08:26:13.171884 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerDied","Data":"f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a"} Nov 22 08:26:14 crc kubenswrapper[4735]: I1122 08:26:14.185391 4735 generic.go:334] "Generic (PLEG): container finished" podID="6c3a99da-447d-4586-b5d8-2214d6df2cbd" containerID="859217d0a75b1a510d7d855ad2228d96c4008c74831dac4111be6e84d7bbd55b" exitCode=0 Nov 22 08:26:14 crc kubenswrapper[4735]: I1122 08:26:14.185498 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" event={"ID":"6c3a99da-447d-4586-b5d8-2214d6df2cbd","Type":"ContainerDied","Data":"859217d0a75b1a510d7d855ad2228d96c4008c74831dac4111be6e84d7bbd55b"} Nov 22 08:26:14 crc kubenswrapper[4735]: I1122 08:26:14.490015 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.190:9292/healthcheck\": read tcp 10.217.0.2:51300->10.217.0.190:9292: read: connection reset by peer" Nov 22 08:26:14 crc kubenswrapper[4735]: I1122 08:26:14.490015 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.190:9292/healthcheck\": read tcp 10.217.0.2:51306->10.217.0.190:9292: read: connection reset by peer" Nov 22 08:26:14 crc kubenswrapper[4735]: I1122 08:26:14.717544 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:26:14 crc kubenswrapper[4735]: I1122 08:26:14.717850 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="14444255-5953-4eef-8a78-46697c99b0e6" containerName="glance-log" containerID="cri-o://2bccddfce55cfbff8b4bf1140d2519a32a053a430f02c542e874e26c3b933d3f" gracePeriod=30 Nov 22 08:26:14 crc kubenswrapper[4735]: I1122 08:26:14.717955 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="14444255-5953-4eef-8a78-46697c99b0e6" containerName="glance-httpd" containerID="cri-o://e3cfe12fee68b54c05ce16495c472a50fc6ff4f401cc61a2daa3370f9b6802a5" gracePeriod=30 Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.108894 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.202218 4735 generic.go:334] "Generic (PLEG): container finished" podID="701597f5-e695-485b-a08c-cec90dad760f" containerID="ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d" exitCode=0 Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.202275 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.202293 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"701597f5-e695-485b-a08c-cec90dad760f","Type":"ContainerDied","Data":"ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d"} Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.202326 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"701597f5-e695-485b-a08c-cec90dad760f","Type":"ContainerDied","Data":"751d33b56e4e45daa63f2dcf6e63d1fe205f89093963f23e7a52db7f4febab2d"} Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.202344 4735 scope.go:117] "RemoveContainer" containerID="ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.206568 4735 generic.go:334] "Generic (PLEG): container finished" podID="14444255-5953-4eef-8a78-46697c99b0e6" containerID="2bccddfce55cfbff8b4bf1140d2519a32a053a430f02c542e874e26c3b933d3f" exitCode=143 Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.206624 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14444255-5953-4eef-8a78-46697c99b0e6","Type":"ContainerDied","Data":"2bccddfce55cfbff8b4bf1140d2519a32a053a430f02c542e874e26c3b933d3f"} Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.245871 4735 scope.go:117] "RemoveContainer" containerID="b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.249275 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-config-data\") pod \"701597f5-e695-485b-a08c-cec90dad760f\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.249363 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-scripts\") pod \"701597f5-e695-485b-a08c-cec90dad760f\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.249434 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl9kn\" (UniqueName: \"kubernetes.io/projected/701597f5-e695-485b-a08c-cec90dad760f-kube-api-access-xl9kn\") pod \"701597f5-e695-485b-a08c-cec90dad760f\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.249546 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-logs\") pod \"701597f5-e695-485b-a08c-cec90dad760f\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.249614 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-httpd-run\") pod \"701597f5-e695-485b-a08c-cec90dad760f\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.249657 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-combined-ca-bundle\") pod \"701597f5-e695-485b-a08c-cec90dad760f\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.249678 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-public-tls-certs\") pod \"701597f5-e695-485b-a08c-cec90dad760f\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.249852 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"701597f5-e695-485b-a08c-cec90dad760f\" (UID: \"701597f5-e695-485b-a08c-cec90dad760f\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.250130 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "701597f5-e695-485b-a08c-cec90dad760f" (UID: "701597f5-e695-485b-a08c-cec90dad760f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.250254 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-logs" (OuterVolumeSpecName: "logs") pod "701597f5-e695-485b-a08c-cec90dad760f" (UID: "701597f5-e695-485b-a08c-cec90dad760f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.250945 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.251042 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/701597f5-e695-485b-a08c-cec90dad760f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.259771 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-scripts" (OuterVolumeSpecName: "scripts") pod "701597f5-e695-485b-a08c-cec90dad760f" (UID: "701597f5-e695-485b-a08c-cec90dad760f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.261447 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/701597f5-e695-485b-a08c-cec90dad760f-kube-api-access-xl9kn" (OuterVolumeSpecName: "kube-api-access-xl9kn") pod "701597f5-e695-485b-a08c-cec90dad760f" (UID: "701597f5-e695-485b-a08c-cec90dad760f"). InnerVolumeSpecName "kube-api-access-xl9kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.297857 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "701597f5-e695-485b-a08c-cec90dad760f" (UID: "701597f5-e695-485b-a08c-cec90dad760f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.322608 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "701597f5-e695-485b-a08c-cec90dad760f" (UID: "701597f5-e695-485b-a08c-cec90dad760f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.322670 4735 scope.go:117] "RemoveContainer" containerID="ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d" Nov 22 08:26:15 crc kubenswrapper[4735]: E1122 08:26:15.324688 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d\": container with ID starting with ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d not found: ID does not exist" containerID="ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.324741 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d"} err="failed to get container status \"ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d\": rpc error: code = NotFound desc = could not find container \"ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d\": container with ID starting with ecdc0a41bfc49c4ae4b57204ad134c6d5c69936c260edad8055a67668bbcd71d not found: ID does not exist" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.324777 4735 scope.go:117] "RemoveContainer" containerID="b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a" Nov 22 08:26:15 crc kubenswrapper[4735]: E1122 08:26:15.325910 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a\": container with ID starting with b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a not found: ID does not exist" containerID="b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.325957 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a"} err="failed to get container status \"b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a\": rpc error: code = NotFound desc = could not find container \"b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a\": container with ID starting with b73d11cf1f183f0179d3a9298c37270f2b314b61da922e9416b673048a84734a not found: ID does not exist" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.361743 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.361992 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.362055 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.362111 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl9kn\" (UniqueName: \"kubernetes.io/projected/701597f5-e695-485b-a08c-cec90dad760f-kube-api-access-xl9kn\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.378544 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "701597f5-e695-485b-a08c-cec90dad760f" (UID: "701597f5-e695-485b-a08c-cec90dad760f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.384586 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-config-data" (OuterVolumeSpecName: "config-data") pod "701597f5-e695-485b-a08c-cec90dad760f" (UID: "701597f5-e695-485b-a08c-cec90dad760f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.397439 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.466869 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.466900 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.466909 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/701597f5-e695-485b-a08c-cec90dad760f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.485991 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.545872 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.569026 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.569839 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8mmv\" (UniqueName: \"kubernetes.io/projected/6c3a99da-447d-4586-b5d8-2214d6df2cbd-kube-api-access-s8mmv\") pod \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.569978 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-scripts\") pod \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.570175 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-config-data\") pod \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.570279 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-combined-ca-bundle\") pod \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\" (UID: \"6c3a99da-447d-4586-b5d8-2214d6df2cbd\") " Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.576757 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c3a99da-447d-4586-b5d8-2214d6df2cbd-kube-api-access-s8mmv" (OuterVolumeSpecName: "kube-api-access-s8mmv") pod "6c3a99da-447d-4586-b5d8-2214d6df2cbd" (UID: "6c3a99da-447d-4586-b5d8-2214d6df2cbd"). InnerVolumeSpecName "kube-api-access-s8mmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.598692 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-scripts" (OuterVolumeSpecName: "scripts") pod "6c3a99da-447d-4586-b5d8-2214d6df2cbd" (UID: "6c3a99da-447d-4586-b5d8-2214d6df2cbd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600040 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:26:15 crc kubenswrapper[4735]: E1122 08:26:15.600589 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" containerName="heat-engine" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600610 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" containerName="heat-engine" Nov 22 08:26:15 crc kubenswrapper[4735]: E1122 08:26:15.600657 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-log" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600664 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-log" Nov 22 08:26:15 crc kubenswrapper[4735]: E1122 08:26:15.600678 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c3a99da-447d-4586-b5d8-2214d6df2cbd" containerName="nova-cell0-conductor-db-sync" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600684 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c3a99da-447d-4586-b5d8-2214d6df2cbd" containerName="nova-cell0-conductor-db-sync" Nov 22 08:26:15 crc kubenswrapper[4735]: E1122 08:26:15.600704 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-httpd" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600711 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-httpd" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600916 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-log" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600937 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="701597f5-e695-485b-a08c-cec90dad760f" containerName="glance-httpd" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600959 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad9eaaf9-ea3b-46e4-a114-dc08e58c0e70" containerName="heat-engine" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.600972 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c3a99da-447d-4586-b5d8-2214d6df2cbd" containerName="nova-cell0-conductor-db-sync" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.604187 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.615256 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.615571 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.615732 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.672839 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.672913 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.673007 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8fdq\" (UniqueName: \"kubernetes.io/projected/2d252eab-255d-49ac-a06a-de9665af67d4-kube-api-access-j8fdq\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.673035 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.673070 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d252eab-255d-49ac-a06a-de9665af67d4-logs\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.673145 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.673241 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.673292 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d252eab-255d-49ac-a06a-de9665af67d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.673361 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8mmv\" (UniqueName: \"kubernetes.io/projected/6c3a99da-447d-4586-b5d8-2214d6df2cbd-kube-api-access-s8mmv\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.673383 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.723674 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c3a99da-447d-4586-b5d8-2214d6df2cbd" (UID: "6c3a99da-447d-4586-b5d8-2214d6df2cbd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.731670 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-config-data" (OuterVolumeSpecName: "config-data") pod "6c3a99da-447d-4586-b5d8-2214d6df2cbd" (UID: "6c3a99da-447d-4586-b5d8-2214d6df2cbd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.784136 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8fdq\" (UniqueName: \"kubernetes.io/projected/2d252eab-255d-49ac-a06a-de9665af67d4-kube-api-access-j8fdq\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.784192 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.784233 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d252eab-255d-49ac-a06a-de9665af67d4-logs\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.784317 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.785924 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.786066 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d252eab-255d-49ac-a06a-de9665af67d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.786101 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.786137 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.786166 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.786402 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d252eab-255d-49ac-a06a-de9665af67d4-logs\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.786467 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.786487 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c3a99da-447d-4586-b5d8-2214d6df2cbd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.791660 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.796892 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.786853 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d252eab-255d-49ac-a06a-de9665af67d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.797512 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.806614 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d252eab-255d-49ac-a06a-de9665af67d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.819734 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8fdq\" (UniqueName: \"kubernetes.io/projected/2d252eab-255d-49ac-a06a-de9665af67d4-kube-api-access-j8fdq\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:15 crc kubenswrapper[4735]: I1122 08:26:15.885263 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"2d252eab-255d-49ac-a06a-de9665af67d4\") " pod="openstack/glance-default-external-api-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.008344 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.242511 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" event={"ID":"6c3a99da-447d-4586-b5d8-2214d6df2cbd","Type":"ContainerDied","Data":"65f2a834c3ff28c4fad2f09341d73811206972b81890195f166cdee947d4b61b"} Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.242901 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65f2a834c3ff28c4fad2f09341d73811206972b81890195f166cdee947d4b61b" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.242993 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sjwjn" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.345547 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.347375 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.351679 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.351867 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tvwv2" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.366303 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.409159 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffa0c54-0cf4-431a-932a-29cc50706cf1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.409764 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx5z5\" (UniqueName: \"kubernetes.io/projected/7ffa0c54-0cf4-431a-932a-29cc50706cf1-kube-api-access-cx5z5\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.409788 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffa0c54-0cf4-431a-932a-29cc50706cf1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.511208 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx5z5\" (UniqueName: \"kubernetes.io/projected/7ffa0c54-0cf4-431a-932a-29cc50706cf1-kube-api-access-cx5z5\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.511255 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffa0c54-0cf4-431a-932a-29cc50706cf1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.511336 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffa0c54-0cf4-431a-932a-29cc50706cf1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.518588 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffa0c54-0cf4-431a-932a-29cc50706cf1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.519376 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffa0c54-0cf4-431a-932a-29cc50706cf1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.528618 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx5z5\" (UniqueName: \"kubernetes.io/projected/7ffa0c54-0cf4-431a-932a-29cc50706cf1-kube-api-access-cx5z5\") pod \"nova-cell0-conductor-0\" (UID: \"7ffa0c54-0cf4-431a-932a-29cc50706cf1\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.689952 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:16 crc kubenswrapper[4735]: I1122 08:26:16.907224 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:26:17 crc kubenswrapper[4735]: I1122 08:26:17.256011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2d252eab-255d-49ac-a06a-de9665af67d4","Type":"ContainerStarted","Data":"e772458d7e79f95bc6946b0b3bd30e4047e72510cf8a3ee08c428821347019a2"} Nov 22 08:26:17 crc kubenswrapper[4735]: I1122 08:26:17.280727 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="701597f5-e695-485b-a08c-cec90dad760f" path="/var/lib/kubelet/pods/701597f5-e695-485b-a08c-cec90dad760f/volumes" Nov 22 08:26:17 crc kubenswrapper[4735]: I1122 08:26:17.379172 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.279369 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7ffa0c54-0cf4-431a-932a-29cc50706cf1","Type":"ContainerStarted","Data":"2cedc67dff9cfa4e89b5b570ada6a7784004e70ff921120fc6895d336e48e3c9"} Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.279940 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7ffa0c54-0cf4-431a-932a-29cc50706cf1","Type":"ContainerStarted","Data":"de458fb1feeb701e5cc7b3687c8b4e7f38ce48df202d718da0303ac82fb685c8"} Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.283230 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.305317 4735 generic.go:334] "Generic (PLEG): container finished" podID="14444255-5953-4eef-8a78-46697c99b0e6" containerID="e3cfe12fee68b54c05ce16495c472a50fc6ff4f401cc61a2daa3370f9b6802a5" exitCode=0 Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.305419 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14444255-5953-4eef-8a78-46697c99b0e6","Type":"ContainerDied","Data":"e3cfe12fee68b54c05ce16495c472a50fc6ff4f401cc61a2daa3370f9b6802a5"} Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.308049 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2d252eab-255d-49ac-a06a-de9665af67d4","Type":"ContainerStarted","Data":"b061ee7a99de8094e735fcd28ce70caa08b6057122828c3c380650fa0620bfd0"} Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.310193 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.310173136 podStartE2EDuration="2.310173136s" podCreationTimestamp="2025-11-22 08:26:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:18.30229154 +0000 UTC m=+1399.906630145" watchObservedRunningTime="2025-11-22 08:26:18.310173136 +0000 UTC m=+1399.914511741" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.526943 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.601092 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-combined-ca-bundle\") pod \"14444255-5953-4eef-8a78-46697c99b0e6\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.601314 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-httpd-run\") pod \"14444255-5953-4eef-8a78-46697c99b0e6\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.601377 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"14444255-5953-4eef-8a78-46697c99b0e6\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.601521 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-config-data\") pod \"14444255-5953-4eef-8a78-46697c99b0e6\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.601594 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9tdb\" (UniqueName: \"kubernetes.io/projected/14444255-5953-4eef-8a78-46697c99b0e6-kube-api-access-g9tdb\") pod \"14444255-5953-4eef-8a78-46697c99b0e6\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.601655 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-internal-tls-certs\") pod \"14444255-5953-4eef-8a78-46697c99b0e6\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.601692 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-logs\") pod \"14444255-5953-4eef-8a78-46697c99b0e6\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.601725 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-scripts\") pod \"14444255-5953-4eef-8a78-46697c99b0e6\" (UID: \"14444255-5953-4eef-8a78-46697c99b0e6\") " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.606343 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "14444255-5953-4eef-8a78-46697c99b0e6" (UID: "14444255-5953-4eef-8a78-46697c99b0e6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.607018 4735 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.610202 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-logs" (OuterVolumeSpecName: "logs") pod "14444255-5953-4eef-8a78-46697c99b0e6" (UID: "14444255-5953-4eef-8a78-46697c99b0e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.617663 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "14444255-5953-4eef-8a78-46697c99b0e6" (UID: "14444255-5953-4eef-8a78-46697c99b0e6"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.617945 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-scripts" (OuterVolumeSpecName: "scripts") pod "14444255-5953-4eef-8a78-46697c99b0e6" (UID: "14444255-5953-4eef-8a78-46697c99b0e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.618151 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14444255-5953-4eef-8a78-46697c99b0e6-kube-api-access-g9tdb" (OuterVolumeSpecName: "kube-api-access-g9tdb") pod "14444255-5953-4eef-8a78-46697c99b0e6" (UID: "14444255-5953-4eef-8a78-46697c99b0e6"). InnerVolumeSpecName "kube-api-access-g9tdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.668931 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14444255-5953-4eef-8a78-46697c99b0e6" (UID: "14444255-5953-4eef-8a78-46697c99b0e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.695859 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-config-data" (OuterVolumeSpecName: "config-data") pod "14444255-5953-4eef-8a78-46697c99b0e6" (UID: "14444255-5953-4eef-8a78-46697c99b0e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.705087 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "14444255-5953-4eef-8a78-46697c99b0e6" (UID: "14444255-5953-4eef-8a78-46697c99b0e6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.709109 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9tdb\" (UniqueName: \"kubernetes.io/projected/14444255-5953-4eef-8a78-46697c99b0e6-kube-api-access-g9tdb\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.709176 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.709192 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14444255-5953-4eef-8a78-46697c99b0e6-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.709206 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.709218 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.709249 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.709263 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14444255-5953-4eef-8a78-46697c99b0e6-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.782834 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 22 08:26:18 crc kubenswrapper[4735]: I1122 08:26:18.811302 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.323614 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14444255-5953-4eef-8a78-46697c99b0e6","Type":"ContainerDied","Data":"99d849f553532428f2f4c54b9c2d8252610eb3ecf967bf805ba7ebf712ca620a"} Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.323934 4735 scope.go:117] "RemoveContainer" containerID="e3cfe12fee68b54c05ce16495c472a50fc6ff4f401cc61a2daa3370f9b6802a5" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.323687 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.327155 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2d252eab-255d-49ac-a06a-de9665af67d4","Type":"ContainerStarted","Data":"d198f76ec89959ab161d812f8fe4f81675d2bc4493521a2f7e48f3ea791fb603"} Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.355749 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.355729924 podStartE2EDuration="4.355729924s" podCreationTimestamp="2025-11-22 08:26:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:19.352716531 +0000 UTC m=+1400.957055146" watchObservedRunningTime="2025-11-22 08:26:19.355729924 +0000 UTC m=+1400.960068529" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.383920 4735 scope.go:117] "RemoveContainer" containerID="2bccddfce55cfbff8b4bf1140d2519a32a053a430f02c542e874e26c3b933d3f" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.392436 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.418652 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.462072 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:26:19 crc kubenswrapper[4735]: E1122 08:26:19.462673 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14444255-5953-4eef-8a78-46697c99b0e6" containerName="glance-httpd" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.462692 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="14444255-5953-4eef-8a78-46697c99b0e6" containerName="glance-httpd" Nov 22 08:26:19 crc kubenswrapper[4735]: E1122 08:26:19.462703 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14444255-5953-4eef-8a78-46697c99b0e6" containerName="glance-log" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.462709 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="14444255-5953-4eef-8a78-46697c99b0e6" containerName="glance-log" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.463203 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="14444255-5953-4eef-8a78-46697c99b0e6" containerName="glance-httpd" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.464372 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="14444255-5953-4eef-8a78-46697c99b0e6" containerName="glance-log" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.472548 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.472657 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.478085 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.480545 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.527918 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78fr6\" (UniqueName: \"kubernetes.io/projected/b9513901-cb1d-4603-9c3f-7fcc2e96f162-kube-api-access-78fr6\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.527963 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9513901-cb1d-4603-9c3f-7fcc2e96f162-logs\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.528010 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.528035 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.528072 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.528109 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b9513901-cb1d-4603-9c3f-7fcc2e96f162-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.528158 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.528181 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.630185 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b9513901-cb1d-4603-9c3f-7fcc2e96f162-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.630285 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.630331 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.630677 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78fr6\" (UniqueName: \"kubernetes.io/projected/b9513901-cb1d-4603-9c3f-7fcc2e96f162-kube-api-access-78fr6\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.630720 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9513901-cb1d-4603-9c3f-7fcc2e96f162-logs\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.630806 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.630837 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.630889 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.631827 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.633598 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b9513901-cb1d-4603-9c3f-7fcc2e96f162-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.633798 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9513901-cb1d-4603-9c3f-7fcc2e96f162-logs\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.641220 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.641819 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.643409 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.643560 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9513901-cb1d-4603-9c3f-7fcc2e96f162-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.653485 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78fr6\" (UniqueName: \"kubernetes.io/projected/b9513901-cb1d-4603-9c3f-7fcc2e96f162-kube-api-access-78fr6\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.670612 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b9513901-cb1d-4603-9c3f-7fcc2e96f162\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:26:19 crc kubenswrapper[4735]: I1122 08:26:19.799130 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:20 crc kubenswrapper[4735]: I1122 08:26:20.402548 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.002633 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.073649 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-config-data\") pod \"05987ef0-923f-428a-a7ea-968d8980c4b0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.073886 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-log-httpd\") pod \"05987ef0-923f-428a-a7ea-968d8980c4b0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.073905 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-sg-core-conf-yaml\") pod \"05987ef0-923f-428a-a7ea-968d8980c4b0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.073920 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-scripts\") pod \"05987ef0-923f-428a-a7ea-968d8980c4b0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.073992 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgfwt\" (UniqueName: \"kubernetes.io/projected/05987ef0-923f-428a-a7ea-968d8980c4b0-kube-api-access-lgfwt\") pod \"05987ef0-923f-428a-a7ea-968d8980c4b0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.074051 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-combined-ca-bundle\") pod \"05987ef0-923f-428a-a7ea-968d8980c4b0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.074099 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-run-httpd\") pod \"05987ef0-923f-428a-a7ea-968d8980c4b0\" (UID: \"05987ef0-923f-428a-a7ea-968d8980c4b0\") " Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.074362 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "05987ef0-923f-428a-a7ea-968d8980c4b0" (UID: "05987ef0-923f-428a-a7ea-968d8980c4b0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.074682 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.075778 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "05987ef0-923f-428a-a7ea-968d8980c4b0" (UID: "05987ef0-923f-428a-a7ea-968d8980c4b0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.084992 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05987ef0-923f-428a-a7ea-968d8980c4b0-kube-api-access-lgfwt" (OuterVolumeSpecName: "kube-api-access-lgfwt") pod "05987ef0-923f-428a-a7ea-968d8980c4b0" (UID: "05987ef0-923f-428a-a7ea-968d8980c4b0"). InnerVolumeSpecName "kube-api-access-lgfwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.103744 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-scripts" (OuterVolumeSpecName: "scripts") pod "05987ef0-923f-428a-a7ea-968d8980c4b0" (UID: "05987ef0-923f-428a-a7ea-968d8980c4b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.134677 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "05987ef0-923f-428a-a7ea-968d8980c4b0" (UID: "05987ef0-923f-428a-a7ea-968d8980c4b0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.176592 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05987ef0-923f-428a-a7ea-968d8980c4b0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.176624 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.176637 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.176648 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgfwt\" (UniqueName: \"kubernetes.io/projected/05987ef0-923f-428a-a7ea-968d8980c4b0-kube-api-access-lgfwt\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.217137 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05987ef0-923f-428a-a7ea-968d8980c4b0" (UID: "05987ef0-923f-428a-a7ea-968d8980c4b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.230266 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-config-data" (OuterVolumeSpecName: "config-data") pod "05987ef0-923f-428a-a7ea-968d8980c4b0" (UID: "05987ef0-923f-428a-a7ea-968d8980c4b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.281853 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.282120 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05987ef0-923f-428a-a7ea-968d8980c4b0-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.291439 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14444255-5953-4eef-8a78-46697c99b0e6" path="/var/lib/kubelet/pods/14444255-5953-4eef-8a78-46697c99b0e6/volumes" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.367507 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9513901-cb1d-4603-9c3f-7fcc2e96f162","Type":"ContainerStarted","Data":"55c40559547aff550560c9366108a4c2e2a8ecccad02fb63e3c7697e89cfd674"} Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.370400 4735 generic.go:334] "Generic (PLEG): container finished" podID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerID="df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e" exitCode=0 Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.370435 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerDied","Data":"df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e"} Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.370453 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"05987ef0-923f-428a-a7ea-968d8980c4b0","Type":"ContainerDied","Data":"e908052ec606f1b79673f0d32d69115b33bb69594b21d31bba4f7a5c41ccef6f"} Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.370487 4735 scope.go:117] "RemoveContainer" containerID="21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.370637 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.418757 4735 scope.go:117] "RemoveContainer" containerID="203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.418761 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.435571 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.450136 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:21 crc kubenswrapper[4735]: E1122 08:26:21.450931 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="sg-core" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.451027 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="sg-core" Nov 22 08:26:21 crc kubenswrapper[4735]: E1122 08:26:21.451118 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="ceilometer-central-agent" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.451242 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="ceilometer-central-agent" Nov 22 08:26:21 crc kubenswrapper[4735]: E1122 08:26:21.451339 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="proxy-httpd" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.451404 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="proxy-httpd" Nov 22 08:26:21 crc kubenswrapper[4735]: E1122 08:26:21.451496 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="ceilometer-notification-agent" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.451566 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="ceilometer-notification-agent" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.452059 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="ceilometer-central-agent" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.452170 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="sg-core" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.452256 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="ceilometer-notification-agent" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.452359 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" containerName="proxy-httpd" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.454493 4735 scope.go:117] "RemoveContainer" containerID="f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.456007 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.459892 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.467364 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.470617 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.535984 4735 scope.go:117] "RemoveContainer" containerID="df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.565845 4735 scope.go:117] "RemoveContainer" containerID="21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57" Nov 22 08:26:21 crc kubenswrapper[4735]: E1122 08:26:21.566221 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57\": container with ID starting with 21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57 not found: ID does not exist" containerID="21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.566248 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57"} err="failed to get container status \"21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57\": rpc error: code = NotFound desc = could not find container \"21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57\": container with ID starting with 21fc660ef3d2814558291a3a9eb799e27aaeabbc7a30d1a98de206d5b3a60e57 not found: ID does not exist" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.566271 4735 scope.go:117] "RemoveContainer" containerID="203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe" Nov 22 08:26:21 crc kubenswrapper[4735]: E1122 08:26:21.566557 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe\": container with ID starting with 203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe not found: ID does not exist" containerID="203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.566577 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe"} err="failed to get container status \"203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe\": rpc error: code = NotFound desc = could not find container \"203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe\": container with ID starting with 203fd3d5a4bf7ba1f6dfbd06d221203fc6494122304bac0410d46ca1aa76c8fe not found: ID does not exist" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.566592 4735 scope.go:117] "RemoveContainer" containerID="f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a" Nov 22 08:26:21 crc kubenswrapper[4735]: E1122 08:26:21.567046 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a\": container with ID starting with f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a not found: ID does not exist" containerID="f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.567075 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a"} err="failed to get container status \"f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a\": rpc error: code = NotFound desc = could not find container \"f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a\": container with ID starting with f55d98138a62b9e8e9effbf4a137dcbe679c2820ded45cabf45cae107f56a87a not found: ID does not exist" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.567093 4735 scope.go:117] "RemoveContainer" containerID="df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e" Nov 22 08:26:21 crc kubenswrapper[4735]: E1122 08:26:21.567369 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e\": container with ID starting with df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e not found: ID does not exist" containerID="df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.567393 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e"} err="failed to get container status \"df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e\": rpc error: code = NotFound desc = could not find container \"df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e\": container with ID starting with df5a9dcee194c9b7d82bfc7689c2fb9e2c73f08078fba4aa9445f2c798f9034e not found: ID does not exist" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.591990 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f4tl\" (UniqueName: \"kubernetes.io/projected/51a54874-7e4e-4764-8434-4d3b5863ccbb-kube-api-access-8f4tl\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.592039 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-config-data\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.592095 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-scripts\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.592149 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.592208 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.592234 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-run-httpd\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.592267 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-log-httpd\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.694889 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-scripts\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.694984 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.695042 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.695070 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-run-httpd\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.695101 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-log-httpd\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.695163 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f4tl\" (UniqueName: \"kubernetes.io/projected/51a54874-7e4e-4764-8434-4d3b5863ccbb-kube-api-access-8f4tl\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.695181 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-config-data\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.698931 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-run-httpd\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.701786 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-scripts\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.701876 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-log-httpd\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.702227 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-config-data\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.705215 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.714308 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.728239 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f4tl\" (UniqueName: \"kubernetes.io/projected/51a54874-7e4e-4764-8434-4d3b5863ccbb-kube-api-access-8f4tl\") pod \"ceilometer-0\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " pod="openstack/ceilometer-0" Nov 22 08:26:21 crc kubenswrapper[4735]: I1122 08:26:21.783861 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:22 crc kubenswrapper[4735]: I1122 08:26:22.316424 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:22 crc kubenswrapper[4735]: W1122 08:26:22.327315 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a54874_7e4e_4764_8434_4d3b5863ccbb.slice/crio-b20a11d5110f73b41da8341575193cb273bb1fa97e486401464fb0d7e3a7ffc5 WatchSource:0}: Error finding container b20a11d5110f73b41da8341575193cb273bb1fa97e486401464fb0d7e3a7ffc5: Status 404 returned error can't find the container with id b20a11d5110f73b41da8341575193cb273bb1fa97e486401464fb0d7e3a7ffc5 Nov 22 08:26:22 crc kubenswrapper[4735]: I1122 08:26:22.381940 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9513901-cb1d-4603-9c3f-7fcc2e96f162","Type":"ContainerStarted","Data":"48d1ba203e199577c4206f8d792014e41aad26701325989770f369cb57f5b321"} Nov 22 08:26:22 crc kubenswrapper[4735]: I1122 08:26:22.381997 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9513901-cb1d-4603-9c3f-7fcc2e96f162","Type":"ContainerStarted","Data":"035a3160e97fda0cc24a591960d3ae506e8ffd53920e5fed43f80ef5bcac163b"} Nov 22 08:26:22 crc kubenswrapper[4735]: I1122 08:26:22.385264 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerStarted","Data":"b20a11d5110f73b41da8341575193cb273bb1fa97e486401464fb0d7e3a7ffc5"} Nov 22 08:26:22 crc kubenswrapper[4735]: I1122 08:26:22.411819 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.411802454 podStartE2EDuration="3.411802454s" podCreationTimestamp="2025-11-22 08:26:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:22.401509752 +0000 UTC m=+1404.005848377" watchObservedRunningTime="2025-11-22 08:26:22.411802454 +0000 UTC m=+1404.016141059" Nov 22 08:26:23 crc kubenswrapper[4735]: I1122 08:26:23.276321 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05987ef0-923f-428a-a7ea-968d8980c4b0" path="/var/lib/kubelet/pods/05987ef0-923f-428a-a7ea-968d8980c4b0/volumes" Nov 22 08:26:23 crc kubenswrapper[4735]: I1122 08:26:23.397140 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerStarted","Data":"fa8c4a774ad811827112e0824ce65639ce720a87e2d80478a8617628c68c6e72"} Nov 22 08:26:24 crc kubenswrapper[4735]: I1122 08:26:24.430979 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerStarted","Data":"4132e5879af0a80610af1d08a948e03fea83e3473f18db3b9b4e749f30607b0d"} Nov 22 08:26:25 crc kubenswrapper[4735]: I1122 08:26:25.442819 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerStarted","Data":"da1dcc30b46d63d9775ff1db4c04468e8e0e46606e13ad764c2394ba02a27bd8"} Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.008913 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.011417 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.042192 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.062600 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.456771 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerStarted","Data":"f5332be9213f3e63cdaef10f2c586e98d4882fbe4b74cf93592ebbb608790b56"} Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.457162 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.457187 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.457336 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.494828 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8726939200000001 podStartE2EDuration="5.494806172s" podCreationTimestamp="2025-11-22 08:26:21 +0000 UTC" firstStartedPulling="2025-11-22 08:26:22.329907774 +0000 UTC m=+1403.934246379" lastFinishedPulling="2025-11-22 08:26:25.952020016 +0000 UTC m=+1407.556358631" observedRunningTime="2025-11-22 08:26:26.48775729 +0000 UTC m=+1408.092095905" watchObservedRunningTime="2025-11-22 08:26:26.494806172 +0000 UTC m=+1408.099144777" Nov 22 08:26:26 crc kubenswrapper[4735]: I1122 08:26:26.725135 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.439339 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-nl8pr"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.441493 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.443843 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.444053 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.521417 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-nl8pr"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.547397 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-config-data\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.547506 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-scripts\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.547626 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.547649 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-697nn\" (UniqueName: \"kubernetes.io/projected/581f275c-c454-46b5-8775-c10a814cfd4c-kube-api-access-697nn\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.652983 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.653037 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-697nn\" (UniqueName: \"kubernetes.io/projected/581f275c-c454-46b5-8775-c10a814cfd4c-kube-api-access-697nn\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.653149 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-config-data\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.653215 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-scripts\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.669181 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-scripts\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.673605 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.682546 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.685105 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.686558 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-config-data\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.689991 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.721038 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-697nn\" (UniqueName: \"kubernetes.io/projected/581f275c-c454-46b5-8775-c10a814cfd4c-kube-api-access-697nn\") pod \"nova-cell0-cell-mapping-nl8pr\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.721101 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.751505 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.753352 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.764315 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.770899 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.798065 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.799819 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.807285 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.843913 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.862058 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zldk\" (UniqueName: \"kubernetes.io/projected/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-kube-api-access-9zldk\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.862104 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-logs\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.862142 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45q9d\" (UniqueName: \"kubernetes.io/projected/29ef616f-a264-4274-a467-40d6d1ad4ed3-kube-api-access-45q9d\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.862169 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ef616f-a264-4274-a467-40d6d1ad4ed3-logs\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.862208 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.862239 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.862299 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-config-data\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.862362 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-config-data\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.885337 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.939685 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6p8hd"] Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.941524 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968317 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-config-data\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968386 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968511 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zldk\" (UniqueName: \"kubernetes.io/projected/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-kube-api-access-9zldk\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968528 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-logs\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968563 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45q9d\" (UniqueName: \"kubernetes.io/projected/29ef616f-a264-4274-a467-40d6d1ad4ed3-kube-api-access-45q9d\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968590 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ef616f-a264-4274-a467-40d6d1ad4ed3-logs\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968625 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968642 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-config-data\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968666 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968691 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpzqv\" (UniqueName: \"kubernetes.io/projected/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-kube-api-access-cpzqv\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.968744 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-config-data\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.970113 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ef616f-a264-4274-a467-40d6d1ad4ed3-logs\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.980993 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-logs\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.984835 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:27 crc kubenswrapper[4735]: I1122 08:26:27.991113 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-config-data\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.002847 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-config-data\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.004417 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.008186 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zldk\" (UniqueName: \"kubernetes.io/projected/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-kube-api-access-9zldk\") pod \"nova-metadata-0\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " pod="openstack/nova-metadata-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.010358 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45q9d\" (UniqueName: \"kubernetes.io/projected/29ef616f-a264-4274-a467-40d6d1ad4ed3-kube-api-access-45q9d\") pod \"nova-api-0\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " pod="openstack/nova-api-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.022879 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.025018 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.032699 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.036530 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6p8hd"] Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.058525 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072145 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072211 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-svc\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072278 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072313 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072345 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-config-data\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072374 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dtkc\" (UniqueName: \"kubernetes.io/projected/58815e92-d627-4845-99df-143d68ba6bd2-kube-api-access-4dtkc\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072398 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpzqv\" (UniqueName: \"kubernetes.io/projected/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-kube-api-access-cpzqv\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072480 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-config\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.072514 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.083971 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.092127 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-config-data\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.098220 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpzqv\" (UniqueName: \"kubernetes.io/projected/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-kube-api-access-cpzqv\") pod \"nova-scheduler-0\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174221 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174333 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174380 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dtkc\" (UniqueName: \"kubernetes.io/projected/58815e92-d627-4845-99df-143d68ba6bd2-kube-api-access-4dtkc\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174443 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174489 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-config\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174585 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174612 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np6wj\" (UniqueName: \"kubernetes.io/projected/72e15695-e11e-4559-b412-bc0a5983bec9-kube-api-access-np6wj\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174631 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-svc\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.174683 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.177361 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-svc\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.178063 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.178686 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.179494 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h7slb"] Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.179773 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-config\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.180298 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.183597 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.201567 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h7slb"] Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.217592 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dtkc\" (UniqueName: \"kubernetes.io/projected/58815e92-d627-4845-99df-143d68ba6bd2-kube-api-access-4dtkc\") pod \"dnsmasq-dns-9b86998b5-6p8hd\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.235322 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.250649 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.280315 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-catalog-content\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.280423 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.280500 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x68qt\" (UniqueName: \"kubernetes.io/projected/ca123e9d-79c1-4c31-b633-59c1ba594be7-kube-api-access-x68qt\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.280676 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np6wj\" (UniqueName: \"kubernetes.io/projected/72e15695-e11e-4559-b412-bc0a5983bec9-kube-api-access-np6wj\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.280745 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-utilities\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.280844 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.286354 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.291520 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.311110 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np6wj\" (UniqueName: \"kubernetes.io/projected/72e15695-e11e-4559-b412-bc0a5983bec9-kube-api-access-np6wj\") pod \"nova-cell1-novncproxy-0\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.354647 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.382915 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x68qt\" (UniqueName: \"kubernetes.io/projected/ca123e9d-79c1-4c31-b633-59c1ba594be7-kube-api-access-x68qt\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.383389 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-utilities\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.383513 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-catalog-content\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.384290 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-catalog-content\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.384529 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.385481 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-utilities\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.405036 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.410976 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x68qt\" (UniqueName: \"kubernetes.io/projected/ca123e9d-79c1-4c31-b633-59c1ba594be7-kube-api-access-x68qt\") pod \"certified-operators-h7slb\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.522310 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:28 crc kubenswrapper[4735]: I1122 08:26:28.669046 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-nl8pr"] Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.089985 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.419495 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.524237 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29ef616f-a264-4274-a467-40d6d1ad4ed3","Type":"ContainerStarted","Data":"1b3ca8c22945cef46d08a8c94b0da78faabdea143a065ba70fd78c46e7d8ad32"} Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.526931 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5","Type":"ContainerStarted","Data":"282bd67c88e4cfbba1026b6b55055fdca3b6a70e8cc46fcca400e0dbe03b7e57"} Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.532443 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nl8pr" event={"ID":"581f275c-c454-46b5-8775-c10a814cfd4c","Type":"ContainerStarted","Data":"ec98ac524eab262f16419899093b6f59c27840dee024c005bfa24c59f8bd51bb"} Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.532498 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nl8pr" event={"ID":"581f275c-c454-46b5-8775-c10a814cfd4c","Type":"ContainerStarted","Data":"42ef92399753aa86110b61a0b8a52abed6061d0d0b5e5efcf971d43544efaa3f"} Nov 22 08:26:29 crc kubenswrapper[4735]: W1122 08:26:29.541591 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3685d5e6_bc4c_4b12_8dab_6c9307b857fb.slice/crio-b5256615f4a4292d9f5929e87512927346def24558511fcb76d48681fd6c78cf WatchSource:0}: Error finding container b5256615f4a4292d9f5929e87512927346def24558511fcb76d48681fd6c78cf: Status 404 returned error can't find the container with id b5256615f4a4292d9f5929e87512927346def24558511fcb76d48681fd6c78cf Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.547785 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.630788 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-nl8pr" podStartSLOduration=2.630766188 podStartE2EDuration="2.630766188s" podCreationTimestamp="2025-11-22 08:26:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:29.562357376 +0000 UTC m=+1411.166695981" watchObservedRunningTime="2025-11-22 08:26:29.630766188 +0000 UTC m=+1411.235104793" Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.637926 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6p8hd"] Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.698104 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.799971 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.800077 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.872156 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h7slb"] Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.882584 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:29 crc kubenswrapper[4735]: W1122 08:26:29.887434 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca123e9d_79c1_4c31_b633_59c1ba594be7.slice/crio-3eff12aa7d5202cd20d8b48cd9fc12c099d34512db5fad1969ecc734cfb3bdec WatchSource:0}: Error finding container 3eff12aa7d5202cd20d8b48cd9fc12c099d34512db5fad1969ecc734cfb3bdec: Status 404 returned error can't find the container with id 3eff12aa7d5202cd20d8b48cd9fc12c099d34512db5fad1969ecc734cfb3bdec Nov 22 08:26:29 crc kubenswrapper[4735]: I1122 08:26:29.903395 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.386612 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5brff"] Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.392172 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.396419 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.396479 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.415113 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5brff"] Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.457272 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-config-data\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.457327 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-scripts\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.457350 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfvch\" (UniqueName: \"kubernetes.io/projected/b3d5067c-9432-415b-a20b-d0822ae72701-kube-api-access-pfvch\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.457795 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.557574 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"72e15695-e11e-4559-b412-bc0a5983bec9","Type":"ContainerStarted","Data":"d7ab54451e96c88c76de29c415e30c350bb8b3f0942a42b6ff8cfa5614a63e0b"} Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.562157 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3685d5e6-bc4c-4b12-8dab-6c9307b857fb","Type":"ContainerStarted","Data":"b5256615f4a4292d9f5929e87512927346def24558511fcb76d48681fd6c78cf"} Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.563596 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-config-data\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.563635 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-scripts\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.563661 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfvch\" (UniqueName: \"kubernetes.io/projected/b3d5067c-9432-415b-a20b-d0822ae72701-kube-api-access-pfvch\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.563766 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.565938 4735 generic.go:334] "Generic (PLEG): container finished" podID="58815e92-d627-4845-99df-143d68ba6bd2" containerID="29db1d9c7a41cc25f47cf5168ada56d7455374ae0964ca0429b4f116e0deba1e" exitCode=0 Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.566043 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" event={"ID":"58815e92-d627-4845-99df-143d68ba6bd2","Type":"ContainerDied","Data":"29db1d9c7a41cc25f47cf5168ada56d7455374ae0964ca0429b4f116e0deba1e"} Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.566073 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" event={"ID":"58815e92-d627-4845-99df-143d68ba6bd2","Type":"ContainerStarted","Data":"140ec08fc8597756902cfe29833e4e6d4cd4c1888927ee1b701341105bd8b3b9"} Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.571153 4735 generic.go:334] "Generic (PLEG): container finished" podID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerID="90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697" exitCode=0 Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.571207 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7slb" event={"ID":"ca123e9d-79c1-4c31-b633-59c1ba594be7","Type":"ContainerDied","Data":"90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697"} Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.571251 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7slb" event={"ID":"ca123e9d-79c1-4c31-b633-59c1ba594be7","Type":"ContainerStarted","Data":"3eff12aa7d5202cd20d8b48cd9fc12c099d34512db5fad1969ecc734cfb3bdec"} Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.572829 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-scripts\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.573147 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.573177 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.573898 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.584167 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-config-data\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.600369 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfvch\" (UniqueName: \"kubernetes.io/projected/b3d5067c-9432-415b-a20b-d0822ae72701-kube-api-access-pfvch\") pod \"nova-cell1-conductor-db-sync-5brff\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:30 crc kubenswrapper[4735]: I1122 08:26:30.737088 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:31 crc kubenswrapper[4735]: I1122 08:26:31.427945 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5brff"] Nov 22 08:26:31 crc kubenswrapper[4735]: W1122 08:26:31.436846 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3d5067c_9432_415b_a20b_d0822ae72701.slice/crio-d5b8e25464839ae985a7048bfbf036a5036b42660f4cc8719f6ef6483117db54 WatchSource:0}: Error finding container d5b8e25464839ae985a7048bfbf036a5036b42660f4cc8719f6ef6483117db54: Status 404 returned error can't find the container with id d5b8e25464839ae985a7048bfbf036a5036b42660f4cc8719f6ef6483117db54 Nov 22 08:26:31 crc kubenswrapper[4735]: I1122 08:26:31.593613 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5brff" event={"ID":"b3d5067c-9432-415b-a20b-d0822ae72701","Type":"ContainerStarted","Data":"d5b8e25464839ae985a7048bfbf036a5036b42660f4cc8719f6ef6483117db54"} Nov 22 08:26:31 crc kubenswrapper[4735]: I1122 08:26:31.630298 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" event={"ID":"58815e92-d627-4845-99df-143d68ba6bd2","Type":"ContainerStarted","Data":"9ffbeb65bba7d6a85c4e62e2a114e21faf6505ccb5783a05f61cf8230dc60f47"} Nov 22 08:26:31 crc kubenswrapper[4735]: I1122 08:26:31.630598 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:31 crc kubenswrapper[4735]: I1122 08:26:31.853115 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" podStartSLOduration=4.853096453 podStartE2EDuration="4.853096453s" podCreationTimestamp="2025-11-22 08:26:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:31.72576095 +0000 UTC m=+1413.330099555" watchObservedRunningTime="2025-11-22 08:26:31.853096453 +0000 UTC m=+1413.457435058" Nov 22 08:26:31 crc kubenswrapper[4735]: I1122 08:26:31.854339 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:26:31 crc kubenswrapper[4735]: I1122 08:26:31.888149 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:32 crc kubenswrapper[4735]: I1122 08:26:32.218475 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 22 08:26:32 crc kubenswrapper[4735]: I1122 08:26:32.218822 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:26:32 crc kubenswrapper[4735]: I1122 08:26:32.222074 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 22 08:26:32 crc kubenswrapper[4735]: I1122 08:26:32.659708 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5brff" event={"ID":"b3d5067c-9432-415b-a20b-d0822ae72701","Type":"ContainerStarted","Data":"d2e9bde4bfd74a00f14234421cb0317ec1f2322ce85e0bd42c486e36ea437819"} Nov 22 08:26:32 crc kubenswrapper[4735]: I1122 08:26:32.663764 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:26:32 crc kubenswrapper[4735]: I1122 08:26:32.663850 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:26:32 crc kubenswrapper[4735]: I1122 08:26:32.662954 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7slb" event={"ID":"ca123e9d-79c1-4c31-b633-59c1ba594be7","Type":"ContainerStarted","Data":"29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc"} Nov 22 08:26:32 crc kubenswrapper[4735]: I1122 08:26:32.691297 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-5brff" podStartSLOduration=2.691276879 podStartE2EDuration="2.691276879s" podCreationTimestamp="2025-11-22 08:26:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:32.679860716 +0000 UTC m=+1414.284199321" watchObservedRunningTime="2025-11-22 08:26:32.691276879 +0000 UTC m=+1414.295615484" Nov 22 08:26:34 crc kubenswrapper[4735]: I1122 08:26:34.696233 4735 generic.go:334] "Generic (PLEG): container finished" podID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerID="29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc" exitCode=0 Nov 22 08:26:34 crc kubenswrapper[4735]: I1122 08:26:34.696272 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7slb" event={"ID":"ca123e9d-79c1-4c31-b633-59c1ba594be7","Type":"ContainerDied","Data":"29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc"} Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.105573 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.106274 4735 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.204341 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.732777 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3685d5e6-bc4c-4b12-8dab-6c9307b857fb","Type":"ContainerStarted","Data":"fdd1048e3216e862621161bed4d5e022e84999a6d9c9ebc1f31f595e38762a2f"} Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.737334 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerName="nova-metadata-log" containerID="cri-o://73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc" gracePeriod=30 Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.737424 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5","Type":"ContainerStarted","Data":"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc"} Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.737447 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5","Type":"ContainerStarted","Data":"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc"} Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.737498 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerName="nova-metadata-metadata" containerID="cri-o://31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc" gracePeriod=30 Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.750905 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.961619367 podStartE2EDuration="9.750885067s" podCreationTimestamp="2025-11-22 08:26:27 +0000 UTC" firstStartedPulling="2025-11-22 08:26:29.559633871 +0000 UTC m=+1411.163972476" lastFinishedPulling="2025-11-22 08:26:35.348899571 +0000 UTC m=+1416.953238176" observedRunningTime="2025-11-22 08:26:36.74622033 +0000 UTC m=+1418.350558945" watchObservedRunningTime="2025-11-22 08:26:36.750885067 +0000 UTC m=+1418.355223672" Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.751146 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29ef616f-a264-4274-a467-40d6d1ad4ed3","Type":"ContainerStarted","Data":"9f8c38306bb03e5eea55b6a87d08d1531265d84a5fe32b8f987f8170ac70bb9d"} Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.751179 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29ef616f-a264-4274-a467-40d6d1ad4ed3","Type":"ContainerStarted","Data":"af4c7ccce6550ef1dc65fdaea882fd8d7b13a6e0453bd58fd250d49197c591f9"} Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.760894 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7slb" event={"ID":"ca123e9d-79c1-4c31-b633-59c1ba594be7","Type":"ContainerStarted","Data":"92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3"} Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.764927 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="72e15695-e11e-4559-b412-bc0a5983bec9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://cdbf0f8270ad299805c0cd3d10637561c328c031eb4d1b62498a410db0d63767" gracePeriod=30 Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.765208 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"72e15695-e11e-4559-b412-bc0a5983bec9","Type":"ContainerStarted","Data":"cdbf0f8270ad299805c0cd3d10637561c328c031eb4d1b62498a410db0d63767"} Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.801837 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.835054668 podStartE2EDuration="9.801810041s" podCreationTimestamp="2025-11-22 08:26:27 +0000 UTC" firstStartedPulling="2025-11-22 08:26:29.377884481 +0000 UTC m=+1410.982223086" lastFinishedPulling="2025-11-22 08:26:35.344639854 +0000 UTC m=+1416.948978459" observedRunningTime="2025-11-22 08:26:36.769072705 +0000 UTC m=+1418.373411330" watchObservedRunningTime="2025-11-22 08:26:36.801810041 +0000 UTC m=+1418.406148646" Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.821376 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=4.168294663 podStartE2EDuration="9.821354526s" podCreationTimestamp="2025-11-22 08:26:27 +0000 UTC" firstStartedPulling="2025-11-22 08:26:29.689743281 +0000 UTC m=+1411.294081886" lastFinishedPulling="2025-11-22 08:26:35.342803134 +0000 UTC m=+1416.947141749" observedRunningTime="2025-11-22 08:26:36.792683541 +0000 UTC m=+1418.397022226" watchObservedRunningTime="2025-11-22 08:26:36.821354526 +0000 UTC m=+1418.425693131" Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.843748 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.587476376 podStartE2EDuration="9.843728958s" podCreationTimestamp="2025-11-22 08:26:27 +0000 UTC" firstStartedPulling="2025-11-22 08:26:29.109658204 +0000 UTC m=+1410.713996809" lastFinishedPulling="2025-11-22 08:26:35.365910786 +0000 UTC m=+1416.970249391" observedRunningTime="2025-11-22 08:26:36.811838295 +0000 UTC m=+1418.416176920" watchObservedRunningTime="2025-11-22 08:26:36.843728958 +0000 UTC m=+1418.448067563" Nov 22 08:26:36 crc kubenswrapper[4735]: I1122 08:26:36.850538 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h7slb" podStartSLOduration=3.7563900180000003 podStartE2EDuration="8.850520853s" podCreationTimestamp="2025-11-22 08:26:28 +0000 UTC" firstStartedPulling="2025-11-22 08:26:30.57703029 +0000 UTC m=+1412.181368895" lastFinishedPulling="2025-11-22 08:26:35.671161135 +0000 UTC m=+1417.275499730" observedRunningTime="2025-11-22 08:26:36.835896593 +0000 UTC m=+1418.440235208" watchObservedRunningTime="2025-11-22 08:26:36.850520853 +0000 UTC m=+1418.454859458" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.517688 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.595312 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-logs\") pod \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.595554 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-config-data\") pod \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.595621 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-combined-ca-bundle\") pod \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.595689 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zldk\" (UniqueName: \"kubernetes.io/projected/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-kube-api-access-9zldk\") pod \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\" (UID: \"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5\") " Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.597278 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-logs" (OuterVolumeSpecName: "logs") pod "f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" (UID: "f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.611435 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-kube-api-access-9zldk" (OuterVolumeSpecName: "kube-api-access-9zldk") pod "f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" (UID: "f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5"). InnerVolumeSpecName "kube-api-access-9zldk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.642749 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" (UID: "f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.648549 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-config-data" (OuterVolumeSpecName: "config-data") pod "f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" (UID: "f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.698786 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.698827 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.698852 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zldk\" (UniqueName: \"kubernetes.io/projected/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-kube-api-access-9zldk\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.698866 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.800729 4735 generic.go:334] "Generic (PLEG): container finished" podID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerID="31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc" exitCode=0 Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.800766 4735 generic.go:334] "Generic (PLEG): container finished" podID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerID="73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc" exitCode=143 Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.800833 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.800825 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5","Type":"ContainerDied","Data":"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc"} Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.800903 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5","Type":"ContainerDied","Data":"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc"} Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.800927 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5","Type":"ContainerDied","Data":"282bd67c88e4cfbba1026b6b55055fdca3b6a70e8cc46fcca400e0dbe03b7e57"} Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.800948 4735 scope.go:117] "RemoveContainer" containerID="31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.890506 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.895632 4735 scope.go:117] "RemoveContainer" containerID="73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.930769 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.959500 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:37 crc kubenswrapper[4735]: E1122 08:26:37.960103 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerName="nova-metadata-log" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.960120 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerName="nova-metadata-log" Nov 22 08:26:37 crc kubenswrapper[4735]: E1122 08:26:37.960161 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerName="nova-metadata-metadata" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.960169 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerName="nova-metadata-metadata" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.960419 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerName="nova-metadata-metadata" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.960437 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" containerName="nova-metadata-log" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.961919 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.965769 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 08:26:37 crc kubenswrapper[4735]: I1122 08:26:37.978808 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.008085 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.008191 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-config-data\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.008216 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.008278 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkms4\" (UniqueName: \"kubernetes.io/projected/7414c30e-a070-48fb-a259-ab4256439dd7-kube-api-access-hkms4\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.008330 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7414c30e-a070-48fb-a259-ab4256439dd7-logs\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.019351 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.035643 4735 scope.go:117] "RemoveContainer" containerID="31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc" Nov 22 08:26:38 crc kubenswrapper[4735]: E1122 08:26:38.050634 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc\": container with ID starting with 31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc not found: ID does not exist" containerID="31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.050675 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc"} err="failed to get container status \"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc\": rpc error: code = NotFound desc = could not find container \"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc\": container with ID starting with 31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc not found: ID does not exist" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.050699 4735 scope.go:117] "RemoveContainer" containerID="73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc" Nov 22 08:26:38 crc kubenswrapper[4735]: E1122 08:26:38.051777 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc\": container with ID starting with 73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc not found: ID does not exist" containerID="73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.051808 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc"} err="failed to get container status \"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc\": rpc error: code = NotFound desc = could not find container \"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc\": container with ID starting with 73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc not found: ID does not exist" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.051875 4735 scope.go:117] "RemoveContainer" containerID="31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.052214 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc"} err="failed to get container status \"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc\": rpc error: code = NotFound desc = could not find container \"31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc\": container with ID starting with 31059fef94892c05cad3d4f83c5cd45f764b39241dd1813e9270436d958db6cc not found: ID does not exist" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.052257 4735 scope.go:117] "RemoveContainer" containerID="73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.052532 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc"} err="failed to get container status \"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc\": rpc error: code = NotFound desc = could not find container \"73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc\": container with ID starting with 73527fe26b83dc6a38d4653b27c8d2916a5369571138242e1131cb079b3bfefc not found: ID does not exist" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.110244 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.110393 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-config-data\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.110428 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.110616 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkms4\" (UniqueName: \"kubernetes.io/projected/7414c30e-a070-48fb-a259-ab4256439dd7-kube-api-access-hkms4\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.110698 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7414c30e-a070-48fb-a259-ab4256439dd7-logs\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.111600 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7414c30e-a070-48fb-a259-ab4256439dd7-logs\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.117078 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-config-data\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.129981 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.135384 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.136218 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkms4\" (UniqueName: \"kubernetes.io/projected/7414c30e-a070-48fb-a259-ab4256439dd7-kube-api-access-hkms4\") pod \"nova-metadata-0\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.236150 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.236510 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.336597 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.358392 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.358439 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.386699 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.405290 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.446808 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.471751 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-lsxsl"] Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.471997 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" podUID="0e32c687-6521-4512-8578-8c52fe8765d4" containerName="dnsmasq-dns" containerID="cri-o://01da5ec42d9c40d97f8ff2d6a29f264b5b3089d8820d4baebaba235ac8adc066" gracePeriod=10 Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.523323 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.523360 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.837974 4735 generic.go:334] "Generic (PLEG): container finished" podID="0e32c687-6521-4512-8578-8c52fe8765d4" containerID="01da5ec42d9c40d97f8ff2d6a29f264b5b3089d8820d4baebaba235ac8adc066" exitCode=0 Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.838057 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" event={"ID":"0e32c687-6521-4512-8578-8c52fe8765d4","Type":"ContainerDied","Data":"01da5ec42d9c40d97f8ff2d6a29f264b5b3089d8820d4baebaba235ac8adc066"} Nov 22 08:26:38 crc kubenswrapper[4735]: I1122 08:26:38.887070 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 22 08:26:39 crc kubenswrapper[4735]: W1122 08:26:39.080547 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7414c30e_a070_48fb_a259_ab4256439dd7.slice/crio-1f450c2a189162b8f43c53308c3a658f16814764c69a2ab6b937a20ced9a65d8 WatchSource:0}: Error finding container 1f450c2a189162b8f43c53308c3a658f16814764c69a2ab6b937a20ced9a65d8: Status 404 returned error can't find the container with id 1f450c2a189162b8f43c53308c3a658f16814764c69a2ab6b937a20ced9a65d8 Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.088208 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.294709 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.297980 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5" path="/var/lib/kubelet/pods/f02d41b4-1ef5-4b5c-bbef-3da2aa6a54a5/volumes" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.320658 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.233:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.320993 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.233:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.456197 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-nb\") pod \"0e32c687-6521-4512-8578-8c52fe8765d4\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.456617 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-swift-storage-0\") pod \"0e32c687-6521-4512-8578-8c52fe8765d4\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.456661 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wv77\" (UniqueName: \"kubernetes.io/projected/0e32c687-6521-4512-8578-8c52fe8765d4-kube-api-access-4wv77\") pod \"0e32c687-6521-4512-8578-8c52fe8765d4\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.456700 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-svc\") pod \"0e32c687-6521-4512-8578-8c52fe8765d4\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.456888 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-sb\") pod \"0e32c687-6521-4512-8578-8c52fe8765d4\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.456927 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-config\") pod \"0e32c687-6521-4512-8578-8c52fe8765d4\" (UID: \"0e32c687-6521-4512-8578-8c52fe8765d4\") " Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.464162 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e32c687-6521-4512-8578-8c52fe8765d4-kube-api-access-4wv77" (OuterVolumeSpecName: "kube-api-access-4wv77") pod "0e32c687-6521-4512-8578-8c52fe8765d4" (UID: "0e32c687-6521-4512-8578-8c52fe8765d4"). InnerVolumeSpecName "kube-api-access-4wv77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.559407 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wv77\" (UniqueName: \"kubernetes.io/projected/0e32c687-6521-4512-8578-8c52fe8765d4-kube-api-access-4wv77\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.560527 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-config" (OuterVolumeSpecName: "config") pod "0e32c687-6521-4512-8578-8c52fe8765d4" (UID: "0e32c687-6521-4512-8578-8c52fe8765d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.619320 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0e32c687-6521-4512-8578-8c52fe8765d4" (UID: "0e32c687-6521-4512-8578-8c52fe8765d4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.621950 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0e32c687-6521-4512-8578-8c52fe8765d4" (UID: "0e32c687-6521-4512-8578-8c52fe8765d4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.622115 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0e32c687-6521-4512-8578-8c52fe8765d4" (UID: "0e32c687-6521-4512-8578-8c52fe8765d4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.625735 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e32c687-6521-4512-8578-8c52fe8765d4" (UID: "0e32c687-6521-4512-8578-8c52fe8765d4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.665147 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-h7slb" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="registry-server" probeResult="failure" output=< Nov 22 08:26:39 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:26:39 crc kubenswrapper[4735]: > Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.673823 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.673863 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.673880 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.673891 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.673902 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e32c687-6521-4512-8578-8c52fe8765d4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.856037 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7414c30e-a070-48fb-a259-ab4256439dd7","Type":"ContainerStarted","Data":"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b"} Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.856109 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7414c30e-a070-48fb-a259-ab4256439dd7","Type":"ContainerStarted","Data":"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c"} Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.856122 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7414c30e-a070-48fb-a259-ab4256439dd7","Type":"ContainerStarted","Data":"1f450c2a189162b8f43c53308c3a658f16814764c69a2ab6b937a20ced9a65d8"} Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.858445 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" event={"ID":"0e32c687-6521-4512-8578-8c52fe8765d4","Type":"ContainerDied","Data":"285d468c0081595a2e413cfad63fb1be48193e7f87b312dc0e6ed98cd03e9f40"} Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.858515 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-lsxsl" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.858544 4735 scope.go:117] "RemoveContainer" containerID="01da5ec42d9c40d97f8ff2d6a29f264b5b3089d8820d4baebaba235ac8adc066" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.900889 4735 scope.go:117] "RemoveContainer" containerID="546ffdf531969359975342136534ed420644c59deb0da3471b75bd6acfa81138" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.903528 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.903500769 podStartE2EDuration="2.903500769s" podCreationTimestamp="2025-11-22 08:26:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:39.879729898 +0000 UTC m=+1421.484068503" watchObservedRunningTime="2025-11-22 08:26:39.903500769 +0000 UTC m=+1421.507839394" Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.932659 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-lsxsl"] Nov 22 08:26:39 crc kubenswrapper[4735]: I1122 08:26:39.948015 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-lsxsl"] Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.286169 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e32c687-6521-4512-8578-8c52fe8765d4" path="/var/lib/kubelet/pods/0e32c687-6521-4512-8578-8c52fe8765d4/volumes" Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.440054 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.440380 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="ceilometer-central-agent" containerID="cri-o://fa8c4a774ad811827112e0824ce65639ce720a87e2d80478a8617628c68c6e72" gracePeriod=30 Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.440603 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="proxy-httpd" containerID="cri-o://f5332be9213f3e63cdaef10f2c586e98d4882fbe4b74cf93592ebbb608790b56" gracePeriod=30 Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.440677 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="sg-core" containerID="cri-o://da1dcc30b46d63d9775ff1db4c04468e8e0e46606e13ad764c2394ba02a27bd8" gracePeriod=30 Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.440741 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="ceilometer-notification-agent" containerID="cri-o://4132e5879af0a80610af1d08a948e03fea83e3473f18db3b9b4e749f30607b0d" gracePeriod=30 Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.455838 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.889353 4735 generic.go:334] "Generic (PLEG): container finished" podID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerID="f5332be9213f3e63cdaef10f2c586e98d4882fbe4b74cf93592ebbb608790b56" exitCode=0 Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.889699 4735 generic.go:334] "Generic (PLEG): container finished" podID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerID="da1dcc30b46d63d9775ff1db4c04468e8e0e46606e13ad764c2394ba02a27bd8" exitCode=2 Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.889440 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerDied","Data":"f5332be9213f3e63cdaef10f2c586e98d4882fbe4b74cf93592ebbb608790b56"} Nov 22 08:26:41 crc kubenswrapper[4735]: I1122 08:26:41.889754 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerDied","Data":"da1dcc30b46d63d9775ff1db4c04468e8e0e46606e13ad764c2394ba02a27bd8"} Nov 22 08:26:42 crc kubenswrapper[4735]: I1122 08:26:42.905943 4735 generic.go:334] "Generic (PLEG): container finished" podID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerID="fa8c4a774ad811827112e0824ce65639ce720a87e2d80478a8617628c68c6e72" exitCode=0 Nov 22 08:26:42 crc kubenswrapper[4735]: I1122 08:26:42.906036 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerDied","Data":"fa8c4a774ad811827112e0824ce65639ce720a87e2d80478a8617628c68c6e72"} Nov 22 08:26:42 crc kubenswrapper[4735]: I1122 08:26:42.909553 4735 generic.go:334] "Generic (PLEG): container finished" podID="581f275c-c454-46b5-8775-c10a814cfd4c" containerID="ec98ac524eab262f16419899093b6f59c27840dee024c005bfa24c59f8bd51bb" exitCode=0 Nov 22 08:26:42 crc kubenswrapper[4735]: I1122 08:26:42.909615 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nl8pr" event={"ID":"581f275c-c454-46b5-8775-c10a814cfd4c","Type":"ContainerDied","Data":"ec98ac524eab262f16419899093b6f59c27840dee024c005bfa24c59f8bd51bb"} Nov 22 08:26:43 crc kubenswrapper[4735]: I1122 08:26:43.337567 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:26:43 crc kubenswrapper[4735]: I1122 08:26:43.337848 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:26:43 crc kubenswrapper[4735]: I1122 08:26:43.920615 4735 generic.go:334] "Generic (PLEG): container finished" podID="b3d5067c-9432-415b-a20b-d0822ae72701" containerID="d2e9bde4bfd74a00f14234421cb0317ec1f2322ce85e0bd42c486e36ea437819" exitCode=0 Nov 22 08:26:43 crc kubenswrapper[4735]: I1122 08:26:43.920706 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5brff" event={"ID":"b3d5067c-9432-415b-a20b-d0822ae72701","Type":"ContainerDied","Data":"d2e9bde4bfd74a00f14234421cb0317ec1f2322ce85e0bd42c486e36ea437819"} Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.421281 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.510590 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-combined-ca-bundle\") pod \"581f275c-c454-46b5-8775-c10a814cfd4c\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.510666 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-scripts\") pod \"581f275c-c454-46b5-8775-c10a814cfd4c\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.510698 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-config-data\") pod \"581f275c-c454-46b5-8775-c10a814cfd4c\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.511193 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-697nn\" (UniqueName: \"kubernetes.io/projected/581f275c-c454-46b5-8775-c10a814cfd4c-kube-api-access-697nn\") pod \"581f275c-c454-46b5-8775-c10a814cfd4c\" (UID: \"581f275c-c454-46b5-8775-c10a814cfd4c\") " Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.556781 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-scripts" (OuterVolumeSpecName: "scripts") pod "581f275c-c454-46b5-8775-c10a814cfd4c" (UID: "581f275c-c454-46b5-8775-c10a814cfd4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.560617 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/581f275c-c454-46b5-8775-c10a814cfd4c-kube-api-access-697nn" (OuterVolumeSpecName: "kube-api-access-697nn") pod "581f275c-c454-46b5-8775-c10a814cfd4c" (UID: "581f275c-c454-46b5-8775-c10a814cfd4c"). InnerVolumeSpecName "kube-api-access-697nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.567717 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "581f275c-c454-46b5-8775-c10a814cfd4c" (UID: "581f275c-c454-46b5-8775-c10a814cfd4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.616684 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-config-data" (OuterVolumeSpecName: "config-data") pod "581f275c-c454-46b5-8775-c10a814cfd4c" (UID: "581f275c-c454-46b5-8775-c10a814cfd4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.620258 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.620281 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.620290 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/581f275c-c454-46b5-8775-c10a814cfd4c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.620298 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-697nn\" (UniqueName: \"kubernetes.io/projected/581f275c-c454-46b5-8775-c10a814cfd4c-kube-api-access-697nn\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.942741 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nl8pr" Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.943564 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nl8pr" event={"ID":"581f275c-c454-46b5-8775-c10a814cfd4c","Type":"ContainerDied","Data":"42ef92399753aa86110b61a0b8a52abed6061d0d0b5e5efcf971d43544efaa3f"} Nov 22 08:26:44 crc kubenswrapper[4735]: I1122 08:26:44.943615 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42ef92399753aa86110b61a0b8a52abed6061d0d0b5e5efcf971d43544efaa3f" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.129393 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.132077 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-log" containerID="cri-o://af4c7ccce6550ef1dc65fdaea882fd8d7b13a6e0453bd58fd250d49197c591f9" gracePeriod=30 Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.132212 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-api" containerID="cri-o://9f8c38306bb03e5eea55b6a87d08d1531265d84a5fe32b8f987f8170ac70bb9d" gracePeriod=30 Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.164592 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.164872 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3685d5e6-bc4c-4b12-8dab-6c9307b857fb" containerName="nova-scheduler-scheduler" containerID="cri-o://fdd1048e3216e862621161bed4d5e022e84999a6d9c9ebc1f31f595e38762a2f" gracePeriod=30 Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.179532 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.179880 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" containerName="nova-metadata-log" containerID="cri-o://123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c" gracePeriod=30 Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.180366 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" containerName="nova-metadata-metadata" containerID="cri-o://1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b" gracePeriod=30 Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.606910 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.647043 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-combined-ca-bundle\") pod \"b3d5067c-9432-415b-a20b-d0822ae72701\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.647153 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-scripts\") pod \"b3d5067c-9432-415b-a20b-d0822ae72701\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.647363 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-config-data\") pod \"b3d5067c-9432-415b-a20b-d0822ae72701\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.647513 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfvch\" (UniqueName: \"kubernetes.io/projected/b3d5067c-9432-415b-a20b-d0822ae72701-kube-api-access-pfvch\") pod \"b3d5067c-9432-415b-a20b-d0822ae72701\" (UID: \"b3d5067c-9432-415b-a20b-d0822ae72701\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.658664 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d5067c-9432-415b-a20b-d0822ae72701-kube-api-access-pfvch" (OuterVolumeSpecName: "kube-api-access-pfvch") pod "b3d5067c-9432-415b-a20b-d0822ae72701" (UID: "b3d5067c-9432-415b-a20b-d0822ae72701"). InnerVolumeSpecName "kube-api-access-pfvch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.669670 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-scripts" (OuterVolumeSpecName: "scripts") pod "b3d5067c-9432-415b-a20b-d0822ae72701" (UID: "b3d5067c-9432-415b-a20b-d0822ae72701"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.712424 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-config-data" (OuterVolumeSpecName: "config-data") pod "b3d5067c-9432-415b-a20b-d0822ae72701" (UID: "b3d5067c-9432-415b-a20b-d0822ae72701"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.743583 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3d5067c-9432-415b-a20b-d0822ae72701" (UID: "b3d5067c-9432-415b-a20b-d0822ae72701"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.749976 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.750008 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.750017 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d5067c-9432-415b-a20b-d0822ae72701-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.750026 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfvch\" (UniqueName: \"kubernetes.io/projected/b3d5067c-9432-415b-a20b-d0822ae72701-kube-api-access-pfvch\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.852252 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.953780 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-nova-metadata-tls-certs\") pod \"7414c30e-a070-48fb-a259-ab4256439dd7\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.960190 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-combined-ca-bundle\") pod \"7414c30e-a070-48fb-a259-ab4256439dd7\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.960296 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-config-data\") pod \"7414c30e-a070-48fb-a259-ab4256439dd7\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.960381 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7414c30e-a070-48fb-a259-ab4256439dd7-logs\") pod \"7414c30e-a070-48fb-a259-ab4256439dd7\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.960490 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkms4\" (UniqueName: \"kubernetes.io/projected/7414c30e-a070-48fb-a259-ab4256439dd7-kube-api-access-hkms4\") pod \"7414c30e-a070-48fb-a259-ab4256439dd7\" (UID: \"7414c30e-a070-48fb-a259-ab4256439dd7\") " Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.961968 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7414c30e-a070-48fb-a259-ab4256439dd7-logs" (OuterVolumeSpecName: "logs") pod "7414c30e-a070-48fb-a259-ab4256439dd7" (UID: "7414c30e-a070-48fb-a259-ab4256439dd7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.977967 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7414c30e-a070-48fb-a259-ab4256439dd7-kube-api-access-hkms4" (OuterVolumeSpecName: "kube-api-access-hkms4") pod "7414c30e-a070-48fb-a259-ab4256439dd7" (UID: "7414c30e-a070-48fb-a259-ab4256439dd7"). InnerVolumeSpecName "kube-api-access-hkms4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.988873 4735 generic.go:334] "Generic (PLEG): container finished" podID="7414c30e-a070-48fb-a259-ab4256439dd7" containerID="1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b" exitCode=0 Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.988914 4735 generic.go:334] "Generic (PLEG): container finished" podID="7414c30e-a070-48fb-a259-ab4256439dd7" containerID="123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c" exitCode=143 Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.988992 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7414c30e-a070-48fb-a259-ab4256439dd7","Type":"ContainerDied","Data":"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b"} Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.989021 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7414c30e-a070-48fb-a259-ab4256439dd7","Type":"ContainerDied","Data":"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c"} Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.989034 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7414c30e-a070-48fb-a259-ab4256439dd7","Type":"ContainerDied","Data":"1f450c2a189162b8f43c53308c3a658f16814764c69a2ab6b937a20ced9a65d8"} Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.989051 4735 scope.go:117] "RemoveContainer" containerID="1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.989271 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.997095 4735 generic.go:334] "Generic (PLEG): container finished" podID="3685d5e6-bc4c-4b12-8dab-6c9307b857fb" containerID="fdd1048e3216e862621161bed4d5e022e84999a6d9c9ebc1f31f595e38762a2f" exitCode=0 Nov 22 08:26:45 crc kubenswrapper[4735]: I1122 08:26:45.997169 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3685d5e6-bc4c-4b12-8dab-6c9307b857fb","Type":"ContainerDied","Data":"fdd1048e3216e862621161bed4d5e022e84999a6d9c9ebc1f31f595e38762a2f"} Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.004541 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7414c30e-a070-48fb-a259-ab4256439dd7" (UID: "7414c30e-a070-48fb-a259-ab4256439dd7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.028445 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-config-data" (OuterVolumeSpecName: "config-data") pod "7414c30e-a070-48fb-a259-ab4256439dd7" (UID: "7414c30e-a070-48fb-a259-ab4256439dd7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.036042 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7414c30e-a070-48fb-a259-ab4256439dd7" (UID: "7414c30e-a070-48fb-a259-ab4256439dd7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.036189 4735 generic.go:334] "Generic (PLEG): container finished" podID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerID="4132e5879af0a80610af1d08a948e03fea83e3473f18db3b9b4e749f30607b0d" exitCode=0 Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.036301 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerDied","Data":"4132e5879af0a80610af1d08a948e03fea83e3473f18db3b9b4e749f30607b0d"} Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.045004 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.052480 4735 generic.go:334] "Generic (PLEG): container finished" podID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerID="af4c7ccce6550ef1dc65fdaea882fd8d7b13a6e0453bd58fd250d49197c591f9" exitCode=143 Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.055916 4735 scope.go:117] "RemoveContainer" containerID="123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.057190 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5brff" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.066101 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" containerName="nova-metadata-metadata" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066141 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" containerName="nova-metadata-metadata" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.066184 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581f275c-c454-46b5-8775-c10a814cfd4c" containerName="nova-manage" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066192 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="581f275c-c454-46b5-8775-c10a814cfd4c" containerName="nova-manage" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.066218 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e32c687-6521-4512-8578-8c52fe8765d4" containerName="init" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066228 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e32c687-6521-4512-8578-8c52fe8765d4" containerName="init" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.066245 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d5067c-9432-415b-a20b-d0822ae72701" containerName="nova-cell1-conductor-db-sync" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066253 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d5067c-9432-415b-a20b-d0822ae72701" containerName="nova-cell1-conductor-db-sync" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.066276 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" containerName="nova-metadata-log" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066284 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" containerName="nova-metadata-log" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.066295 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e32c687-6521-4512-8578-8c52fe8765d4" containerName="dnsmasq-dns" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066303 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e32c687-6521-4512-8578-8c52fe8765d4" containerName="dnsmasq-dns" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066595 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d5067c-9432-415b-a20b-d0822ae72701" containerName="nova-cell1-conductor-db-sync" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066618 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" containerName="nova-metadata-log" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066638 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" containerName="nova-metadata-metadata" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066654 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e32c687-6521-4512-8578-8c52fe8765d4" containerName="dnsmasq-dns" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.066672 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="581f275c-c454-46b5-8775-c10a814cfd4c" containerName="nova-manage" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.070966 4735 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.071021 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.071036 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7414c30e-a070-48fb-a259-ab4256439dd7-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.071048 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7414c30e-a070-48fb-a259-ab4256439dd7-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.071060 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkms4\" (UniqueName: \"kubernetes.io/projected/7414c30e-a070-48fb-a259-ab4256439dd7-kube-api-access-hkms4\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.072050 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29ef616f-a264-4274-a467-40d6d1ad4ed3","Type":"ContainerDied","Data":"af4c7ccce6550ef1dc65fdaea882fd8d7b13a6e0453bd58fd250d49197c591f9"} Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.072095 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.072115 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5brff" event={"ID":"b3d5067c-9432-415b-a20b-d0822ae72701","Type":"ContainerDied","Data":"d5b8e25464839ae985a7048bfbf036a5036b42660f4cc8719f6ef6483117db54"} Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.072130 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5b8e25464839ae985a7048bfbf036a5036b42660f4cc8719f6ef6483117db54" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.072212 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.085901 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.122142 4735 scope.go:117] "RemoveContainer" containerID="1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.123899 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b\": container with ID starting with 1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b not found: ID does not exist" containerID="1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.123939 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b"} err="failed to get container status \"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b\": rpc error: code = NotFound desc = could not find container \"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b\": container with ID starting with 1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b not found: ID does not exist" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.123974 4735 scope.go:117] "RemoveContainer" containerID="123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.124391 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c\": container with ID starting with 123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c not found: ID does not exist" containerID="123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.124443 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c"} err="failed to get container status \"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c\": rpc error: code = NotFound desc = could not find container \"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c\": container with ID starting with 123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c not found: ID does not exist" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.124490 4735 scope.go:117] "RemoveContainer" containerID="1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.124861 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b"} err="failed to get container status \"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b\": rpc error: code = NotFound desc = could not find container \"1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b\": container with ID starting with 1bed7994b3f428cfef954136e3e08ccbafad19f6c50967d3da7a6ec1c9f5623b not found: ID does not exist" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.124889 4735 scope.go:117] "RemoveContainer" containerID="123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.125259 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c"} err="failed to get container status \"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c\": rpc error: code = NotFound desc = could not find container \"123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c\": container with ID starting with 123754bf034aa94bf685f017696a9d1161540fc672e352215f8d54ccaa3c236c not found: ID does not exist" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.172647 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.172712 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.172808 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dffnz\" (UniqueName: \"kubernetes.io/projected/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-kube-api-access-dffnz\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.275512 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.275594 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.275730 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dffnz\" (UniqueName: \"kubernetes.io/projected/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-kube-api-access-dffnz\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.280552 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.292658 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.302039 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dffnz\" (UniqueName: \"kubernetes.io/projected/b04e0725-3b2f-411c-aeaa-1c9d586bbe37-kube-api-access-dffnz\") pod \"nova-cell1-conductor-0\" (UID: \"b04e0725-3b2f-411c-aeaa-1c9d586bbe37\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.306644 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.379029 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-config-data\") pod \"51a54874-7e4e-4764-8434-4d3b5863ccbb\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.381267 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-combined-ca-bundle\") pod \"51a54874-7e4e-4764-8434-4d3b5863ccbb\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.381416 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f4tl\" (UniqueName: \"kubernetes.io/projected/51a54874-7e4e-4764-8434-4d3b5863ccbb-kube-api-access-8f4tl\") pod \"51a54874-7e4e-4764-8434-4d3b5863ccbb\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.381544 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-sg-core-conf-yaml\") pod \"51a54874-7e4e-4764-8434-4d3b5863ccbb\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.381633 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-log-httpd\") pod \"51a54874-7e4e-4764-8434-4d3b5863ccbb\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.381664 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-scripts\") pod \"51a54874-7e4e-4764-8434-4d3b5863ccbb\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.381749 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-run-httpd\") pod \"51a54874-7e4e-4764-8434-4d3b5863ccbb\" (UID: \"51a54874-7e4e-4764-8434-4d3b5863ccbb\") " Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.383326 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "51a54874-7e4e-4764-8434-4d3b5863ccbb" (UID: "51a54874-7e4e-4764-8434-4d3b5863ccbb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.384211 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "51a54874-7e4e-4764-8434-4d3b5863ccbb" (UID: "51a54874-7e4e-4764-8434-4d3b5863ccbb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.392428 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-scripts" (OuterVolumeSpecName: "scripts") pod "51a54874-7e4e-4764-8434-4d3b5863ccbb" (UID: "51a54874-7e4e-4764-8434-4d3b5863ccbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.394852 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a54874-7e4e-4764-8434-4d3b5863ccbb-kube-api-access-8f4tl" (OuterVolumeSpecName: "kube-api-access-8f4tl") pod "51a54874-7e4e-4764-8434-4d3b5863ccbb" (UID: "51a54874-7e4e-4764-8434-4d3b5863ccbb"). InnerVolumeSpecName "kube-api-access-8f4tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.418121 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "51a54874-7e4e-4764-8434-4d3b5863ccbb" (UID: "51a54874-7e4e-4764-8434-4d3b5863ccbb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.423804 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.486224 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f4tl\" (UniqueName: \"kubernetes.io/projected/51a54874-7e4e-4764-8434-4d3b5863ccbb-kube-api-access-8f4tl\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.486259 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.486272 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.486284 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.486295 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51a54874-7e4e-4764-8434-4d3b5863ccbb-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.499159 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51a54874-7e4e-4764-8434-4d3b5863ccbb" (UID: "51a54874-7e4e-4764-8434-4d3b5863ccbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.511275 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-config-data" (OuterVolumeSpecName: "config-data") pod "51a54874-7e4e-4764-8434-4d3b5863ccbb" (UID: "51a54874-7e4e-4764-8434-4d3b5863ccbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.589700 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.589739 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a54874-7e4e-4764-8434-4d3b5863ccbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.746631 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.762596 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.779774 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.780405 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="proxy-httpd" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.780420 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="proxy-httpd" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.780450 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="ceilometer-central-agent" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.780480 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="ceilometer-central-agent" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.780521 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="ceilometer-notification-agent" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.780530 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="ceilometer-notification-agent" Nov 22 08:26:46 crc kubenswrapper[4735]: E1122 08:26:46.780544 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="sg-core" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.780551 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="sg-core" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.780836 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="ceilometer-central-agent" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.780863 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="sg-core" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.780876 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="ceilometer-notification-agent" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.780891 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" containerName="proxy-httpd" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.782514 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.785676 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.786480 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.791762 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.797697 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhql7\" (UniqueName: \"kubernetes.io/projected/049fb0cb-f818-4234-a5ee-3bbfb518aec8-kube-api-access-rhql7\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.798283 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-config-data\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.798349 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.798430 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/049fb0cb-f818-4234-a5ee-3bbfb518aec8-logs\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.798586 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.900826 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.900929 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/049fb0cb-f818-4234-a5ee-3bbfb518aec8-logs\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.900970 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.901063 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhql7\" (UniqueName: \"kubernetes.io/projected/049fb0cb-f818-4234-a5ee-3bbfb518aec8-kube-api-access-rhql7\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.901273 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-config-data\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.901767 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/049fb0cb-f818-4234-a5ee-3bbfb518aec8-logs\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.913377 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.913497 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-config-data\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.913759 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:46 crc kubenswrapper[4735]: I1122 08:26:46.921103 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhql7\" (UniqueName: \"kubernetes.io/projected/049fb0cb-f818-4234-a5ee-3bbfb518aec8-kube-api-access-rhql7\") pod \"nova-metadata-0\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " pod="openstack/nova-metadata-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.069853 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51a54874-7e4e-4764-8434-4d3b5863ccbb","Type":"ContainerDied","Data":"b20a11d5110f73b41da8341575193cb273bb1fa97e486401464fb0d7e3a7ffc5"} Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.069908 4735 scope.go:117] "RemoveContainer" containerID="f5332be9213f3e63cdaef10f2c586e98d4882fbe4b74cf93592ebbb608790b56" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.070106 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.084741 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7bdb732-0307-475f-8224-9c5e3a3e66a6" containerID="f469819db7cac1e2222c8ce9694ba77c7f27cafc3ab708ff5b94553b1a7f114a" exitCode=137 Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.084800 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-99b96cbf9-4rfhb" event={"ID":"b7bdb732-0307-475f-8224-9c5e3a3e66a6","Type":"ContainerDied","Data":"f469819db7cac1e2222c8ce9694ba77c7f27cafc3ab708ff5b94553b1a7f114a"} Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.105854 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.119839 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.112657 4735 scope.go:117] "RemoveContainer" containerID="da1dcc30b46d63d9775ff1db4c04468e8e0e46606e13ad764c2394ba02a27bd8" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.144834 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.164716 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.181120 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.191306 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.195969 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.196064 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.214195 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.214527 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cndhr\" (UniqueName: \"kubernetes.io/projected/a0c5ca73-9b7a-473f-b505-a20ba547e71b-kube-api-access-cndhr\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.214644 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-log-httpd\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.214672 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.214714 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-run-httpd\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.214884 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-scripts\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.214965 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-config-data\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.215477 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.240615 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:47 crc kubenswrapper[4735]: E1122 08:26:47.241681 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-cndhr log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="a0c5ca73-9b7a-473f-b505-a20ba547e71b" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.265821 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.282019 4735 scope.go:117] "RemoveContainer" containerID="4132e5879af0a80610af1d08a948e03fea83e3473f18db3b9b4e749f30607b0d" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.310308 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a54874-7e4e-4764-8434-4d3b5863ccbb" path="/var/lib/kubelet/pods/51a54874-7e4e-4764-8434-4d3b5863ccbb/volumes" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.311265 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7414c30e-a070-48fb-a259-ab4256439dd7" path="/var/lib/kubelet/pods/7414c30e-a070-48fb-a259-ab4256439dd7/volumes" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.316334 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpzqv\" (UniqueName: \"kubernetes.io/projected/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-kube-api-access-cpzqv\") pod \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.316499 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-combined-ca-bundle\") pod \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.316716 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-config-data\") pod \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\" (UID: \"3685d5e6-bc4c-4b12-8dab-6c9307b857fb\") " Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.317060 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-log-httpd\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.317088 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.317126 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-run-httpd\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.317225 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-scripts\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.317280 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-config-data\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.317334 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.317364 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cndhr\" (UniqueName: \"kubernetes.io/projected/a0c5ca73-9b7a-473f-b505-a20ba547e71b-kube-api-access-cndhr\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.318321 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-log-httpd\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.318351 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-run-httpd\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.324727 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-kube-api-access-cpzqv" (OuterVolumeSpecName: "kube-api-access-cpzqv") pod "3685d5e6-bc4c-4b12-8dab-6c9307b857fb" (UID: "3685d5e6-bc4c-4b12-8dab-6c9307b857fb"). InnerVolumeSpecName "kube-api-access-cpzqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.339506 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-scripts\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.339970 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.342892 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-config-data\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.347211 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cndhr\" (UniqueName: \"kubernetes.io/projected/a0c5ca73-9b7a-473f-b505-a20ba547e71b-kube-api-access-cndhr\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.368843 4735 scope.go:117] "RemoveContainer" containerID="fa8c4a774ad811827112e0824ce65639ce720a87e2d80478a8617628c68c6e72" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.375530 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " pod="openstack/ceilometer-0" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.402059 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3685d5e6-bc4c-4b12-8dab-6c9307b857fb" (UID: "3685d5e6-bc4c-4b12-8dab-6c9307b857fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.402168 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-config-data" (OuterVolumeSpecName: "config-data") pod "3685d5e6-bc4c-4b12-8dab-6c9307b857fb" (UID: "3685d5e6-bc4c-4b12-8dab-6c9307b857fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.422102 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.422142 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpzqv\" (UniqueName: \"kubernetes.io/projected/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-kube-api-access-cpzqv\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.422158 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3685d5e6-bc4c-4b12-8dab-6c9307b857fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.604896 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.630994 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-combined-ca-bundle\") pod \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.631180 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zrnn\" (UniqueName: \"kubernetes.io/projected/b7bdb732-0307-475f-8224-9c5e3a3e66a6-kube-api-access-2zrnn\") pod \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.631220 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data-custom\") pod \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.631272 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data\") pod \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\" (UID: \"b7bdb732-0307-475f-8224-9c5e3a3e66a6\") " Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.637052 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b7bdb732-0307-475f-8224-9c5e3a3e66a6" (UID: "b7bdb732-0307-475f-8224-9c5e3a3e66a6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.638790 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7bdb732-0307-475f-8224-9c5e3a3e66a6-kube-api-access-2zrnn" (OuterVolumeSpecName: "kube-api-access-2zrnn") pod "b7bdb732-0307-475f-8224-9c5e3a3e66a6" (UID: "b7bdb732-0307-475f-8224-9c5e3a3e66a6"). InnerVolumeSpecName "kube-api-access-2zrnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.733489 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zrnn\" (UniqueName: \"kubernetes.io/projected/b7bdb732-0307-475f-8224-9c5e3a3e66a6-kube-api-access-2zrnn\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.733518 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.757983 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7bdb732-0307-475f-8224-9c5e3a3e66a6" (UID: "b7bdb732-0307-475f-8224-9c5e3a3e66a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.815387 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.835070 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.850752 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data" (OuterVolumeSpecName: "config-data") pod "b7bdb732-0307-475f-8224-9c5e3a3e66a6" (UID: "b7bdb732-0307-475f-8224-9c5e3a3e66a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:47 crc kubenswrapper[4735]: I1122 08:26:47.937820 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7bdb732-0307-475f-8224-9c5e3a3e66a6-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.097298 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"049fb0cb-f818-4234-a5ee-3bbfb518aec8","Type":"ContainerStarted","Data":"89b3b94cd5481689d2dddc2cc71c04cfe5f2522a83b5b0fe0ed91963d4c3c50d"} Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.099987 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-99b96cbf9-4rfhb" event={"ID":"b7bdb732-0307-475f-8224-9c5e3a3e66a6","Type":"ContainerDied","Data":"afb1fac74cb4c069efe5152227942ccebf7a10dc86d153f133914a06857c789b"} Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.100020 4735 scope.go:117] "RemoveContainer" containerID="f469819db7cac1e2222c8ce9694ba77c7f27cafc3ab708ff5b94553b1a7f114a" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.100170 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-99b96cbf9-4rfhb" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.107279 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3685d5e6-bc4c-4b12-8dab-6c9307b857fb","Type":"ContainerDied","Data":"b5256615f4a4292d9f5929e87512927346def24558511fcb76d48681fd6c78cf"} Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.107290 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.115804 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.115867 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b04e0725-3b2f-411c-aeaa-1c9d586bbe37","Type":"ContainerStarted","Data":"decd24a7ca0ad552ea324796684628bf72b6bb9ac12da56decf8acfbea824c41"} Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.115921 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b04e0725-3b2f-411c-aeaa-1c9d586bbe37","Type":"ContainerStarted","Data":"b611ae5075dfd201e5b1fb6aba19f0a6fce0219e8dda8e9e30e7457dacf09bb0"} Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.116604 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.135181 4735 scope.go:117] "RemoveContainer" containerID="fdd1048e3216e862621161bed4d5e022e84999a6d9c9ebc1f31f595e38762a2f" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.164229 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.164208217 podStartE2EDuration="2.164208217s" podCreationTimestamp="2025-11-22 08:26:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:48.157542294 +0000 UTC m=+1429.761880899" watchObservedRunningTime="2025-11-22 08:26:48.164208217 +0000 UTC m=+1429.768546822" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.182836 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.189635 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.208617 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.218518 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:48 crc kubenswrapper[4735]: E1122 08:26:48.219139 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7bdb732-0307-475f-8224-9c5e3a3e66a6" containerName="heat-api" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.219162 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7bdb732-0307-475f-8224-9c5e3a3e66a6" containerName="heat-api" Nov 22 08:26:48 crc kubenswrapper[4735]: E1122 08:26:48.219205 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3685d5e6-bc4c-4b12-8dab-6c9307b857fb" containerName="nova-scheduler-scheduler" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.219218 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3685d5e6-bc4c-4b12-8dab-6c9307b857fb" containerName="nova-scheduler-scheduler" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.219499 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7bdb732-0307-475f-8224-9c5e3a3e66a6" containerName="heat-api" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.219528 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3685d5e6-bc4c-4b12-8dab-6c9307b857fb" containerName="nova-scheduler-scheduler" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.220484 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.226999 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.241740 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-99b96cbf9-4rfhb"] Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.243588 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-scripts\") pod \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.243645 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-combined-ca-bundle\") pod \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.243676 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-log-httpd\") pod \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.243741 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cndhr\" (UniqueName: \"kubernetes.io/projected/a0c5ca73-9b7a-473f-b505-a20ba547e71b-kube-api-access-cndhr\") pod \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.243774 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-run-httpd\") pod \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.243806 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-config-data\") pod \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.243876 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-sg-core-conf-yaml\") pod \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\" (UID: \"a0c5ca73-9b7a-473f-b505-a20ba547e71b\") " Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.244122 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.244166 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-config-data\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.244295 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pv4p\" (UniqueName: \"kubernetes.io/projected/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-kube-api-access-2pv4p\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.244817 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a0c5ca73-9b7a-473f-b505-a20ba547e71b" (UID: "a0c5ca73-9b7a-473f-b505-a20ba547e71b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.245069 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a0c5ca73-9b7a-473f-b505-a20ba547e71b" (UID: "a0c5ca73-9b7a-473f-b505-a20ba547e71b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.254156 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0c5ca73-9b7a-473f-b505-a20ba547e71b" (UID: "a0c5ca73-9b7a-473f-b505-a20ba547e71b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.254298 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c5ca73-9b7a-473f-b505-a20ba547e71b-kube-api-access-cndhr" (OuterVolumeSpecName: "kube-api-access-cndhr") pod "a0c5ca73-9b7a-473f-b505-a20ba547e71b" (UID: "a0c5ca73-9b7a-473f-b505-a20ba547e71b"). InnerVolumeSpecName "kube-api-access-cndhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.254381 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-scripts" (OuterVolumeSpecName: "scripts") pod "a0c5ca73-9b7a-473f-b505-a20ba547e71b" (UID: "a0c5ca73-9b7a-473f-b505-a20ba547e71b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.256872 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a0c5ca73-9b7a-473f-b505-a20ba547e71b" (UID: "a0c5ca73-9b7a-473f-b505-a20ba547e71b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.258571 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-config-data" (OuterVolumeSpecName: "config-data") pod "a0c5ca73-9b7a-473f-b505-a20ba547e71b" (UID: "a0c5ca73-9b7a-473f-b505-a20ba547e71b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.264328 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-99b96cbf9-4rfhb"] Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.276500 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.348999 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pv4p\" (UniqueName: \"kubernetes.io/projected/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-kube-api-access-2pv4p\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.349226 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.349268 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-config-data\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.350824 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.350851 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.350867 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.350878 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cndhr\" (UniqueName: \"kubernetes.io/projected/a0c5ca73-9b7a-473f-b505-a20ba547e71b-kube-api-access-cndhr\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.350889 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0c5ca73-9b7a-473f-b505-a20ba547e71b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.350899 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.350909 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0c5ca73-9b7a-473f-b505-a20ba547e71b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.355386 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-config-data\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.355426 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.369218 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pv4p\" (UniqueName: \"kubernetes.io/projected/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-kube-api-access-2pv4p\") pod \"nova-scheduler-0\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " pod="openstack/nova-scheduler-0" Nov 22 08:26:48 crc kubenswrapper[4735]: I1122 08:26:48.570191 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.175437 4735 generic.go:334] "Generic (PLEG): container finished" podID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerID="9f8c38306bb03e5eea55b6a87d08d1531265d84a5fe32b8f987f8170ac70bb9d" exitCode=0 Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.175497 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29ef616f-a264-4274-a467-40d6d1ad4ed3","Type":"ContainerDied","Data":"9f8c38306bb03e5eea55b6a87d08d1531265d84a5fe32b8f987f8170ac70bb9d"} Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.179903 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"049fb0cb-f818-4234-a5ee-3bbfb518aec8","Type":"ContainerStarted","Data":"65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798"} Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.179937 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"049fb0cb-f818-4234-a5ee-3bbfb518aec8","Type":"ContainerStarted","Data":"7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7"} Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.194442 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.214707 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.214673039 podStartE2EDuration="3.214673039s" podCreationTimestamp="2025-11-22 08:26:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:49.198081015 +0000 UTC m=+1430.802419630" watchObservedRunningTime="2025-11-22 08:26:49.214673039 +0000 UTC m=+1430.819011644" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.288429 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3685d5e6-bc4c-4b12-8dab-6c9307b857fb" path="/var/lib/kubelet/pods/3685d5e6-bc4c-4b12-8dab-6c9307b857fb/volumes" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.292533 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7bdb732-0307-475f-8224-9c5e3a3e66a6" path="/var/lib/kubelet/pods/b7bdb732-0307-475f-8224-9c5e3a3e66a6/volumes" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.293158 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.297361 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.310849 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.316692 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.328694 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.328952 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.329360 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.479140 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-log-httpd\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.479206 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-run-httpd\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.479362 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.479429 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.479542 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z26b\" (UniqueName: \"kubernetes.io/projected/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-kube-api-access-8z26b\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.479665 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-config-data\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.479685 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-scripts\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.581306 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-scripts\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.581554 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-log-httpd\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.581601 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-run-httpd\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.581629 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.581677 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.581724 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z26b\" (UniqueName: \"kubernetes.io/projected/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-kube-api-access-8z26b\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.581779 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-config-data\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.582753 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-run-httpd\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.583057 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-log-httpd\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.587675 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.588349 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-config-data\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.588996 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-scripts\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.594322 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.596890 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-h7slb" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="registry-server" probeResult="failure" output=< Nov 22 08:26:49 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:26:49 crc kubenswrapper[4735]: > Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.599792 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z26b\" (UniqueName: \"kubernetes.io/projected/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-kube-api-access-8z26b\") pod \"ceilometer-0\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " pod="openstack/ceilometer-0" Nov 22 08:26:49 crc kubenswrapper[4735]: I1122 08:26:49.692197 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.023392 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:26:50 crc kubenswrapper[4735]: W1122 08:26:50.061670 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1a4cf71_e395_4e1c_9ba7_1a3b37c99b35.slice/crio-bc3dd1d60faae1185885083a29c196c9cb3c6ada8fff017d2e566e301f2ce115 WatchSource:0}: Error finding container bc3dd1d60faae1185885083a29c196c9cb3c6ada8fff017d2e566e301f2ce115: Status 404 returned error can't find the container with id bc3dd1d60faae1185885083a29c196c9cb3c6ada8fff017d2e566e301f2ce115 Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.218569 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35","Type":"ContainerStarted","Data":"bc3dd1d60faae1185885083a29c196c9cb3c6ada8fff017d2e566e301f2ce115"} Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.224736 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.297757 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-combined-ca-bundle\") pod \"29ef616f-a264-4274-a467-40d6d1ad4ed3\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.297850 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ef616f-a264-4274-a467-40d6d1ad4ed3-logs\") pod \"29ef616f-a264-4274-a467-40d6d1ad4ed3\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.297876 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-config-data\") pod \"29ef616f-a264-4274-a467-40d6d1ad4ed3\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.298027 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45q9d\" (UniqueName: \"kubernetes.io/projected/29ef616f-a264-4274-a467-40d6d1ad4ed3-kube-api-access-45q9d\") pod \"29ef616f-a264-4274-a467-40d6d1ad4ed3\" (UID: \"29ef616f-a264-4274-a467-40d6d1ad4ed3\") " Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.299788 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29ef616f-a264-4274-a467-40d6d1ad4ed3-logs" (OuterVolumeSpecName: "logs") pod "29ef616f-a264-4274-a467-40d6d1ad4ed3" (UID: "29ef616f-a264-4274-a467-40d6d1ad4ed3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.304412 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29ef616f-a264-4274-a467-40d6d1ad4ed3-kube-api-access-45q9d" (OuterVolumeSpecName: "kube-api-access-45q9d") pod "29ef616f-a264-4274-a467-40d6d1ad4ed3" (UID: "29ef616f-a264-4274-a467-40d6d1ad4ed3"). InnerVolumeSpecName "kube-api-access-45q9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.316070 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.351708 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-config-data" (OuterVolumeSpecName: "config-data") pod "29ef616f-a264-4274-a467-40d6d1ad4ed3" (UID: "29ef616f-a264-4274-a467-40d6d1ad4ed3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.367320 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29ef616f-a264-4274-a467-40d6d1ad4ed3" (UID: "29ef616f-a264-4274-a467-40d6d1ad4ed3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.401476 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.401691 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ef616f-a264-4274-a467-40d6d1ad4ed3-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.401771 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ef616f-a264-4274-a467-40d6d1ad4ed3-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:50 crc kubenswrapper[4735]: I1122 08:26:50.401948 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45q9d\" (UniqueName: \"kubernetes.io/projected/29ef616f-a264-4274-a467-40d6d1ad4ed3-kube-api-access-45q9d\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.238681 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerStarted","Data":"c3f1413acf03125c23f14a0669138f9e868c3bb37dd90b6159ae0674994cd5f7"} Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.238733 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerStarted","Data":"4a9291bd350bf65d69a5d25deae03b8d351f5202c0833faaf932245ed5612360"} Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.243340 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29ef616f-a264-4274-a467-40d6d1ad4ed3","Type":"ContainerDied","Data":"1b3ca8c22945cef46d08a8c94b0da78faabdea143a065ba70fd78c46e7d8ad32"} Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.243384 4735 scope.go:117] "RemoveContainer" containerID="9f8c38306bb03e5eea55b6a87d08d1531265d84a5fe32b8f987f8170ac70bb9d" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.243584 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.294209 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0c5ca73-9b7a-473f-b505-a20ba547e71b" path="/var/lib/kubelet/pods/a0c5ca73-9b7a-473f-b505-a20ba547e71b/volumes" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.295064 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35","Type":"ContainerStarted","Data":"98a9e0b2236cc1e46be7cdc73cf7a000baf80113d8acec7d59ec71d8bb4b5acf"} Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.296648 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.306163 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.308018 4735 scope.go:117] "RemoveContainer" containerID="af4c7ccce6550ef1dc65fdaea882fd8d7b13a6e0453bd58fd250d49197c591f9" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.326541 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:51 crc kubenswrapper[4735]: E1122 08:26:51.327054 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-api" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.327076 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-api" Nov 22 08:26:51 crc kubenswrapper[4735]: E1122 08:26:51.327096 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-log" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.327103 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-log" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.327379 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-api" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.327397 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" containerName="nova-api-log" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.328703 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.334290 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.337345 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.337316827 podStartE2EDuration="3.337316827s" podCreationTimestamp="2025-11-22 08:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:51.315180471 +0000 UTC m=+1432.919519076" watchObservedRunningTime="2025-11-22 08:26:51.337316827 +0000 UTC m=+1432.941655432" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.368076 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.429953 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.430038 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c6rk\" (UniqueName: \"kubernetes.io/projected/8c53b788-2694-4f84-8ad5-e1f1b17b531d-kube-api-access-9c6rk\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.430071 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-config-data\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.430152 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53b788-2694-4f84-8ad5-e1f1b17b531d-logs\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.532128 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.532184 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c6rk\" (UniqueName: \"kubernetes.io/projected/8c53b788-2694-4f84-8ad5-e1f1b17b531d-kube-api-access-9c6rk\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.532207 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-config-data\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.532277 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53b788-2694-4f84-8ad5-e1f1b17b531d-logs\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.532746 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53b788-2694-4f84-8ad5-e1f1b17b531d-logs\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.537415 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-config-data\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.549087 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.568152 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c6rk\" (UniqueName: \"kubernetes.io/projected/8c53b788-2694-4f84-8ad5-e1f1b17b531d-kube-api-access-9c6rk\") pod \"nova-api-0\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " pod="openstack/nova-api-0" Nov 22 08:26:51 crc kubenswrapper[4735]: I1122 08:26:51.660143 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.106823 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.107272 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.148778 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.291787 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c53b788-2694-4f84-8ad5-e1f1b17b531d","Type":"ContainerStarted","Data":"186f87c04a76d7262577ad0a3a95822ec2be38dd96c39e27de25e74bdcda48e2"} Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.295143 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerStarted","Data":"4e60d4add92bb9b9b00d433eaeef48d6d48a9732b6001d661db7ec53e457daaf"} Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.732821 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-mlbnx"] Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.734992 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.749440 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-8ebf-account-create-6ptmj"] Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.751778 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.754414 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.787549 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-mlbnx"] Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.810965 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-8ebf-account-create-6ptmj"] Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.876710 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f63f2a55-15ff-41f3-834f-537c0d9556fb-operator-scripts\") pod \"aodh-db-create-mlbnx\" (UID: \"f63f2a55-15ff-41f3-834f-537c0d9556fb\") " pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.876806 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlwt7\" (UniqueName: \"kubernetes.io/projected/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-kube-api-access-hlwt7\") pod \"aodh-8ebf-account-create-6ptmj\" (UID: \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\") " pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.877014 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h76j8\" (UniqueName: \"kubernetes.io/projected/f63f2a55-15ff-41f3-834f-537c0d9556fb-kube-api-access-h76j8\") pod \"aodh-db-create-mlbnx\" (UID: \"f63f2a55-15ff-41f3-834f-537c0d9556fb\") " pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.877051 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-operator-scripts\") pod \"aodh-8ebf-account-create-6ptmj\" (UID: \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\") " pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.979697 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h76j8\" (UniqueName: \"kubernetes.io/projected/f63f2a55-15ff-41f3-834f-537c0d9556fb-kube-api-access-h76j8\") pod \"aodh-db-create-mlbnx\" (UID: \"f63f2a55-15ff-41f3-834f-537c0d9556fb\") " pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.980134 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-operator-scripts\") pod \"aodh-8ebf-account-create-6ptmj\" (UID: \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\") " pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.980212 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f63f2a55-15ff-41f3-834f-537c0d9556fb-operator-scripts\") pod \"aodh-db-create-mlbnx\" (UID: \"f63f2a55-15ff-41f3-834f-537c0d9556fb\") " pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.980338 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlwt7\" (UniqueName: \"kubernetes.io/projected/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-kube-api-access-hlwt7\") pod \"aodh-8ebf-account-create-6ptmj\" (UID: \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\") " pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.982090 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-operator-scripts\") pod \"aodh-8ebf-account-create-6ptmj\" (UID: \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\") " pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:52 crc kubenswrapper[4735]: I1122 08:26:52.982758 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f63f2a55-15ff-41f3-834f-537c0d9556fb-operator-scripts\") pod \"aodh-db-create-mlbnx\" (UID: \"f63f2a55-15ff-41f3-834f-537c0d9556fb\") " pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.009000 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlwt7\" (UniqueName: \"kubernetes.io/projected/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-kube-api-access-hlwt7\") pod \"aodh-8ebf-account-create-6ptmj\" (UID: \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\") " pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.015046 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h76j8\" (UniqueName: \"kubernetes.io/projected/f63f2a55-15ff-41f3-834f-537c0d9556fb-kube-api-access-h76j8\") pod \"aodh-db-create-mlbnx\" (UID: \"f63f2a55-15ff-41f3-834f-537c0d9556fb\") " pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.076830 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.091891 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.275623 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29ef616f-a264-4274-a467-40d6d1ad4ed3" path="/var/lib/kubelet/pods/29ef616f-a264-4274-a467-40d6d1ad4ed3/volumes" Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.315995 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c53b788-2694-4f84-8ad5-e1f1b17b531d","Type":"ContainerStarted","Data":"7188d585575f843c8658f5c4893b2526cbead83edfc8591c27a190f37987d44d"} Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.316034 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c53b788-2694-4f84-8ad5-e1f1b17b531d","Type":"ContainerStarted","Data":"3ea3e49962d94d8965c1601e2214a5f0749d4f79ae16eae908bbe0250c79ec01"} Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.345335 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.34530907 podStartE2EDuration="2.34530907s" podCreationTimestamp="2025-11-22 08:26:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:53.340960672 +0000 UTC m=+1434.945299357" watchObservedRunningTime="2025-11-22 08:26:53.34530907 +0000 UTC m=+1434.949647695" Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.574069 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.866579 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-8ebf-account-create-6ptmj"] Nov 22 08:26:53 crc kubenswrapper[4735]: I1122 08:26:53.936098 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-mlbnx"] Nov 22 08:26:54 crc kubenswrapper[4735]: I1122 08:26:54.328800 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerStarted","Data":"e733c893683860b8b01e0a8d7c386a05a3e7e3d454f9372451a397057d27eb03"} Nov 22 08:26:54 crc kubenswrapper[4735]: I1122 08:26:54.330073 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8ebf-account-create-6ptmj" event={"ID":"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090","Type":"ContainerStarted","Data":"627f34b0078843df7c995ea540213274c097e4779b8cae6447f11f9dfc287f23"} Nov 22 08:26:54 crc kubenswrapper[4735]: I1122 08:26:54.330105 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8ebf-account-create-6ptmj" event={"ID":"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090","Type":"ContainerStarted","Data":"043f8d373101b63e1b00e2ed8e9e811bf92330bc60c33b0d425fff98f1581f92"} Nov 22 08:26:54 crc kubenswrapper[4735]: I1122 08:26:54.331439 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-mlbnx" event={"ID":"f63f2a55-15ff-41f3-834f-537c0d9556fb","Type":"ContainerStarted","Data":"bb479f3f4433946fd75ead2b2be992fe15d075627bc518838bbd77131e3c843e"} Nov 22 08:26:54 crc kubenswrapper[4735]: I1122 08:26:54.331511 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-mlbnx" event={"ID":"f63f2a55-15ff-41f3-834f-537c0d9556fb","Type":"ContainerStarted","Data":"d59711198ca87ea734c4d8f92209e0b64f4ad60442e8a26290e6b33910cefbff"} Nov 22 08:26:54 crc kubenswrapper[4735]: I1122 08:26:54.348329 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-8ebf-account-create-6ptmj" podStartSLOduration=2.348310194 podStartE2EDuration="2.348310194s" podCreationTimestamp="2025-11-22 08:26:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:54.341437357 +0000 UTC m=+1435.945775962" watchObservedRunningTime="2025-11-22 08:26:54.348310194 +0000 UTC m=+1435.952648799" Nov 22 08:26:54 crc kubenswrapper[4735]: I1122 08:26:54.366551 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-mlbnx" podStartSLOduration=2.366527933 podStartE2EDuration="2.366527933s" podCreationTimestamp="2025-11-22 08:26:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:26:54.354170164 +0000 UTC m=+1435.958508769" watchObservedRunningTime="2025-11-22 08:26:54.366527933 +0000 UTC m=+1435.970866528" Nov 22 08:26:55 crc kubenswrapper[4735]: I1122 08:26:55.349549 4735 generic.go:334] "Generic (PLEG): container finished" podID="ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090" containerID="627f34b0078843df7c995ea540213274c097e4779b8cae6447f11f9dfc287f23" exitCode=0 Nov 22 08:26:55 crc kubenswrapper[4735]: I1122 08:26:55.349871 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8ebf-account-create-6ptmj" event={"ID":"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090","Type":"ContainerDied","Data":"627f34b0078843df7c995ea540213274c097e4779b8cae6447f11f9dfc287f23"} Nov 22 08:26:55 crc kubenswrapper[4735]: I1122 08:26:55.353050 4735 generic.go:334] "Generic (PLEG): container finished" podID="f63f2a55-15ff-41f3-834f-537c0d9556fb" containerID="bb479f3f4433946fd75ead2b2be992fe15d075627bc518838bbd77131e3c843e" exitCode=0 Nov 22 08:26:55 crc kubenswrapper[4735]: I1122 08:26:55.353100 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-mlbnx" event={"ID":"f63f2a55-15ff-41f3-834f-537c0d9556fb","Type":"ContainerDied","Data":"bb479f3f4433946fd75ead2b2be992fe15d075627bc518838bbd77131e3c843e"} Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.373733 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerStarted","Data":"f85c38b3353bd1397b04cfd53a2500e92a3f2ed267c4f25015dbadbdb86e6110"} Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.418264 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.429057426 podStartE2EDuration="7.418243691s" podCreationTimestamp="2025-11-22 08:26:49 +0000 UTC" firstStartedPulling="2025-11-22 08:26:50.32979823 +0000 UTC m=+1431.934136835" lastFinishedPulling="2025-11-22 08:26:55.318984475 +0000 UTC m=+1436.923323100" observedRunningTime="2025-11-22 08:26:56.401718289 +0000 UTC m=+1438.006056924" watchObservedRunningTime="2025-11-22 08:26:56.418243691 +0000 UTC m=+1438.022582316" Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.461769 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.936536 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.943673 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.980324 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-operator-scripts\") pod \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\" (UID: \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\") " Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.980431 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlwt7\" (UniqueName: \"kubernetes.io/projected/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-kube-api-access-hlwt7\") pod \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\" (UID: \"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090\") " Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.980520 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f63f2a55-15ff-41f3-834f-537c0d9556fb-operator-scripts\") pod \"f63f2a55-15ff-41f3-834f-537c0d9556fb\" (UID: \"f63f2a55-15ff-41f3-834f-537c0d9556fb\") " Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.980588 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h76j8\" (UniqueName: \"kubernetes.io/projected/f63f2a55-15ff-41f3-834f-537c0d9556fb-kube-api-access-h76j8\") pod \"f63f2a55-15ff-41f3-834f-537c0d9556fb\" (UID: \"f63f2a55-15ff-41f3-834f-537c0d9556fb\") " Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.980963 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090" (UID: "ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.981033 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f63f2a55-15ff-41f3-834f-537c0d9556fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f63f2a55-15ff-41f3-834f-537c0d9556fb" (UID: "f63f2a55-15ff-41f3-834f-537c0d9556fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.996221 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-kube-api-access-hlwt7" (OuterVolumeSpecName: "kube-api-access-hlwt7") pod "ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090" (UID: "ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090"). InnerVolumeSpecName "kube-api-access-hlwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:56 crc kubenswrapper[4735]: I1122 08:26:56.998411 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f63f2a55-15ff-41f3-834f-537c0d9556fb-kube-api-access-h76j8" (OuterVolumeSpecName: "kube-api-access-h76j8") pod "f63f2a55-15ff-41f3-834f-537c0d9556fb" (UID: "f63f2a55-15ff-41f3-834f-537c0d9556fb"). InnerVolumeSpecName "kube-api-access-h76j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.083631 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h76j8\" (UniqueName: \"kubernetes.io/projected/f63f2a55-15ff-41f3-834f-537c0d9556fb-kube-api-access-h76j8\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.084004 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.084014 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlwt7\" (UniqueName: \"kubernetes.io/projected/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090-kube-api-access-hlwt7\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.084025 4735 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f63f2a55-15ff-41f3-834f-537c0d9556fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.107409 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.107486 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.386992 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8ebf-account-create-6ptmj" event={"ID":"ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090","Type":"ContainerDied","Data":"043f8d373101b63e1b00e2ed8e9e811bf92330bc60c33b0d425fff98f1581f92"} Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.387031 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="043f8d373101b63e1b00e2ed8e9e811bf92330bc60c33b0d425fff98f1581f92" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.387064 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8ebf-account-create-6ptmj" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.388679 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-mlbnx" event={"ID":"f63f2a55-15ff-41f3-834f-537c0d9556fb","Type":"ContainerDied","Data":"d59711198ca87ea734c4d8f92209e0b64f4ad60442e8a26290e6b33910cefbff"} Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.388701 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d59711198ca87ea734c4d8f92209e0b64f4ad60442e8a26290e6b33910cefbff" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.388741 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-mlbnx" Nov 22 08:26:57 crc kubenswrapper[4735]: I1122 08:26:57.388924 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:26:58 crc kubenswrapper[4735]: I1122 08:26:58.131838 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.242:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:26:58 crc kubenswrapper[4735]: I1122 08:26:58.132070 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.242:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:26:58 crc kubenswrapper[4735]: I1122 08:26:58.572124 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 22 08:26:58 crc kubenswrapper[4735]: I1122 08:26:58.611807 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 22 08:26:59 crc kubenswrapper[4735]: I1122 08:26:59.439066 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 22 08:26:59 crc kubenswrapper[4735]: I1122 08:26:59.614413 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-h7slb" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="registry-server" probeResult="failure" output=< Nov 22 08:26:59 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:26:59 crc kubenswrapper[4735]: > Nov 22 08:27:01 crc kubenswrapper[4735]: I1122 08:27:01.661268 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:27:01 crc kubenswrapper[4735]: I1122 08:27:01.661554 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:27:02 crc kubenswrapper[4735]: I1122 08:27:02.743876 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.246:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:27:02 crc kubenswrapper[4735]: I1122 08:27:02.744740 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.246:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.117018 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-rkfmf"] Nov 22 08:27:03 crc kubenswrapper[4735]: E1122 08:27:03.117700 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f63f2a55-15ff-41f3-834f-537c0d9556fb" containerName="mariadb-database-create" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.117723 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="f63f2a55-15ff-41f3-834f-537c0d9556fb" containerName="mariadb-database-create" Nov 22 08:27:03 crc kubenswrapper[4735]: E1122 08:27:03.117779 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090" containerName="mariadb-account-create" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.117788 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090" containerName="mariadb-account-create" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.118890 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090" containerName="mariadb-account-create" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.118929 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="f63f2a55-15ff-41f3-834f-537c0d9556fb" containerName="mariadb-database-create" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.120137 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.125809 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-bfgs7" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.126507 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.127291 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.130921 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.169548 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-rkfmf"] Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.277759 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdztc\" (UniqueName: \"kubernetes.io/projected/4179aca1-9f3a-4cd5-a9ad-3a301a313668-kube-api-access-xdztc\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.278176 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-combined-ca-bundle\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.278281 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-config-data\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.278353 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-scripts\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.380112 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-combined-ca-bundle\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.380170 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-config-data\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.380193 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-scripts\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.380225 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdztc\" (UniqueName: \"kubernetes.io/projected/4179aca1-9f3a-4cd5-a9ad-3a301a313668-kube-api-access-xdztc\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.387028 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-scripts\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.387450 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-config-data\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.389052 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-combined-ca-bundle\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.401408 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdztc\" (UniqueName: \"kubernetes.io/projected/4179aca1-9f3a-4cd5-a9ad-3a301a313668-kube-api-access-xdztc\") pod \"aodh-db-sync-rkfmf\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:03 crc kubenswrapper[4735]: I1122 08:27:03.464883 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:04 crc kubenswrapper[4735]: I1122 08:27:03.999209 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-rkfmf"] Nov 22 08:27:04 crc kubenswrapper[4735]: W1122 08:27:03.999366 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4179aca1_9f3a_4cd5_a9ad_3a301a313668.slice/crio-09f29b8972850cdd0b6d19e4174fae1c64ed70a977f6a4513a01fbbec135e502 WatchSource:0}: Error finding container 09f29b8972850cdd0b6d19e4174fae1c64ed70a977f6a4513a01fbbec135e502: Status 404 returned error can't find the container with id 09f29b8972850cdd0b6d19e4174fae1c64ed70a977f6a4513a01fbbec135e502 Nov 22 08:27:04 crc kubenswrapper[4735]: I1122 08:27:04.495207 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-rkfmf" event={"ID":"4179aca1-9f3a-4cd5-a9ad-3a301a313668","Type":"ContainerStarted","Data":"09f29b8972850cdd0b6d19e4174fae1c64ed70a977f6a4513a01fbbec135e502"} Nov 22 08:27:07 crc kubenswrapper[4735]: I1122 08:27:07.110904 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 08:27:07 crc kubenswrapper[4735]: I1122 08:27:07.111795 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 08:27:07 crc kubenswrapper[4735]: I1122 08:27:07.117053 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 08:27:07 crc kubenswrapper[4735]: I1122 08:27:07.120115 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 08:27:07 crc kubenswrapper[4735]: I1122 08:27:07.559920 4735 generic.go:334] "Generic (PLEG): container finished" podID="72e15695-e11e-4559-b412-bc0a5983bec9" containerID="cdbf0f8270ad299805c0cd3d10637561c328c031eb4d1b62498a410db0d63767" exitCode=137 Nov 22 08:27:07 crc kubenswrapper[4735]: I1122 08:27:07.560004 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"72e15695-e11e-4559-b412-bc0a5983bec9","Type":"ContainerDied","Data":"cdbf0f8270ad299805c0cd3d10637561c328c031eb4d1b62498a410db0d63767"} Nov 22 08:27:07 crc kubenswrapper[4735]: I1122 08:27:07.891846 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.020273 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-combined-ca-bundle\") pod \"72e15695-e11e-4559-b412-bc0a5983bec9\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.020874 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-config-data\") pod \"72e15695-e11e-4559-b412-bc0a5983bec9\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.020960 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np6wj\" (UniqueName: \"kubernetes.io/projected/72e15695-e11e-4559-b412-bc0a5983bec9-kube-api-access-np6wj\") pod \"72e15695-e11e-4559-b412-bc0a5983bec9\" (UID: \"72e15695-e11e-4559-b412-bc0a5983bec9\") " Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.042351 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e15695-e11e-4559-b412-bc0a5983bec9-kube-api-access-np6wj" (OuterVolumeSpecName: "kube-api-access-np6wj") pod "72e15695-e11e-4559-b412-bc0a5983bec9" (UID: "72e15695-e11e-4559-b412-bc0a5983bec9"). InnerVolumeSpecName "kube-api-access-np6wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.074041 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-config-data" (OuterVolumeSpecName: "config-data") pod "72e15695-e11e-4559-b412-bc0a5983bec9" (UID: "72e15695-e11e-4559-b412-bc0a5983bec9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.075875 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72e15695-e11e-4559-b412-bc0a5983bec9" (UID: "72e15695-e11e-4559-b412-bc0a5983bec9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.123814 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.123872 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np6wj\" (UniqueName: \"kubernetes.io/projected/72e15695-e11e-4559-b412-bc0a5983bec9-kube-api-access-np6wj\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.123894 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e15695-e11e-4559-b412-bc0a5983bec9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.573620 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.573615 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"72e15695-e11e-4559-b412-bc0a5983bec9","Type":"ContainerDied","Data":"d7ab54451e96c88c76de29c415e30c350bb8b3f0942a42b6ff8cfa5614a63e0b"} Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.574009 4735 scope.go:117] "RemoveContainer" containerID="cdbf0f8270ad299805c0cd3d10637561c328c031eb4d1b62498a410db0d63767" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.671403 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.690572 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.705325 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:27:08 crc kubenswrapper[4735]: E1122 08:27:08.706041 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e15695-e11e-4559-b412-bc0a5983bec9" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.706124 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e15695-e11e-4559-b412-bc0a5983bec9" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.706476 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e15695-e11e-4559-b412-bc0a5983bec9" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.708150 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.710880 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.711305 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.712263 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.719891 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.850947 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.851070 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94qcj\" (UniqueName: \"kubernetes.io/projected/939a7da6-6cb9-4e53-8875-357317b90738-kube-api-access-94qcj\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.851138 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.851178 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.851204 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.952970 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.953197 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.953452 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.953582 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94qcj\" (UniqueName: \"kubernetes.io/projected/939a7da6-6cb9-4e53-8875-357317b90738-kube-api-access-94qcj\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.954052 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.960100 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.964179 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.966258 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.975295 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/939a7da6-6cb9-4e53-8875-357317b90738-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:08 crc kubenswrapper[4735]: I1122 08:27:08.981891 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94qcj\" (UniqueName: \"kubernetes.io/projected/939a7da6-6cb9-4e53-8875-357317b90738-kube-api-access-94qcj\") pod \"nova-cell1-novncproxy-0\" (UID: \"939a7da6-6cb9-4e53-8875-357317b90738\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:09 crc kubenswrapper[4735]: I1122 08:27:09.037710 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:09 crc kubenswrapper[4735]: I1122 08:27:09.286096 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e15695-e11e-4559-b412-bc0a5983bec9" path="/var/lib/kubelet/pods/72e15695-e11e-4559-b412-bc0a5983bec9/volumes" Nov 22 08:27:09 crc kubenswrapper[4735]: I1122 08:27:09.583425 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-h7slb" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="registry-server" probeResult="failure" output=< Nov 22 08:27:09 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:27:09 crc kubenswrapper[4735]: > Nov 22 08:27:11 crc kubenswrapper[4735]: I1122 08:27:11.664391 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 08:27:11 crc kubenswrapper[4735]: I1122 08:27:11.665430 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 08:27:11 crc kubenswrapper[4735]: I1122 08:27:11.665720 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 08:27:11 crc kubenswrapper[4735]: I1122 08:27:11.668518 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.274750 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.618586 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.622041 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.780624 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh"] Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.783830 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.813624 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh"] Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.861632 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-config\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.862003 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.862613 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.862653 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtc6s\" (UniqueName: \"kubernetes.io/projected/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-kube-api-access-wtc6s\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.862679 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.863122 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.965257 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.965312 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-config\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.965376 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.965423 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.965450 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtc6s\" (UniqueName: \"kubernetes.io/projected/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-kube-api-access-wtc6s\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.965483 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.966224 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.966764 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.968762 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.970689 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-config\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.970708 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:12 crc kubenswrapper[4735]: I1122 08:27:12.984370 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtc6s\" (UniqueName: \"kubernetes.io/projected/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-kube-api-access-wtc6s\") pod \"dnsmasq-dns-6b7bbf7cf9-cxzfh\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:13 crc kubenswrapper[4735]: I1122 08:27:13.114960 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:13 crc kubenswrapper[4735]: I1122 08:27:13.631040 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-rkfmf" event={"ID":"4179aca1-9f3a-4cd5-a9ad-3a301a313668","Type":"ContainerStarted","Data":"0b072957e6f7373a5320c3e2a92f347bb471db0944c469c4e92d077ccc93a5e0"} Nov 22 08:27:13 crc kubenswrapper[4735]: I1122 08:27:13.634898 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"939a7da6-6cb9-4e53-8875-357317b90738","Type":"ContainerStarted","Data":"d301489ebe06d9ccb8111f51b8d6d8635b98ec6c10cf195af5b7996f824cfd1f"} Nov 22 08:27:13 crc kubenswrapper[4735]: I1122 08:27:13.634935 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"939a7da6-6cb9-4e53-8875-357317b90738","Type":"ContainerStarted","Data":"d1b90fc0236dd82b9e625f354f78feb9eca47ec4e0ffe001bb794940424efe3f"} Nov 22 08:27:13 crc kubenswrapper[4735]: I1122 08:27:13.651360 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-rkfmf" podStartSLOduration=1.6927039860000002 podStartE2EDuration="10.651339713s" podCreationTimestamp="2025-11-22 08:27:03 +0000 UTC" firstStartedPulling="2025-11-22 08:27:04.00588348 +0000 UTC m=+1445.610222085" lastFinishedPulling="2025-11-22 08:27:12.964519217 +0000 UTC m=+1454.568857812" observedRunningTime="2025-11-22 08:27:13.648564417 +0000 UTC m=+1455.252903042" watchObservedRunningTime="2025-11-22 08:27:13.651339713 +0000 UTC m=+1455.255678318" Nov 22 08:27:13 crc kubenswrapper[4735]: I1122 08:27:13.669781 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=5.669760116 podStartE2EDuration="5.669760116s" podCreationTimestamp="2025-11-22 08:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:27:13.667432382 +0000 UTC m=+1455.271770987" watchObservedRunningTime="2025-11-22 08:27:13.669760116 +0000 UTC m=+1455.274098721" Nov 22 08:27:13 crc kubenswrapper[4735]: I1122 08:27:13.715267 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh"] Nov 22 08:27:14 crc kubenswrapper[4735]: I1122 08:27:14.038671 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:14 crc kubenswrapper[4735]: I1122 08:27:14.644543 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" event={"ID":"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb","Type":"ContainerStarted","Data":"b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703"} Nov 22 08:27:14 crc kubenswrapper[4735]: I1122 08:27:14.644585 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" event={"ID":"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb","Type":"ContainerStarted","Data":"873066076a75a58f6233f3bcbf80bfa222d501f0a6013e0e1390eb9877d3884a"} Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.429636 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.430356 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="ceilometer-central-agent" containerID="cri-o://c3f1413acf03125c23f14a0669138f9e868c3bb37dd90b6159ae0674994cd5f7" gracePeriod=30 Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.430926 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="ceilometer-notification-agent" containerID="cri-o://4e60d4add92bb9b9b00d433eaeef48d6d48a9732b6001d661db7ec53e457daaf" gracePeriod=30 Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.430977 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="proxy-httpd" containerID="cri-o://f85c38b3353bd1397b04cfd53a2500e92a3f2ed267c4f25015dbadbdb86e6110" gracePeriod=30 Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.430931 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="sg-core" containerID="cri-o://e733c893683860b8b01e0a8d7c386a05a3e7e3d454f9372451a397057d27eb03" gracePeriod=30 Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.450815 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.245:3000/\": EOF" Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.657172 4735 generic.go:334] "Generic (PLEG): container finished" podID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" containerID="b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703" exitCode=0 Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.657283 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" event={"ID":"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb","Type":"ContainerDied","Data":"b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703"} Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.664625 4735 generic.go:334] "Generic (PLEG): container finished" podID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerID="e733c893683860b8b01e0a8d7c386a05a3e7e3d454f9372451a397057d27eb03" exitCode=2 Nov 22 08:27:15 crc kubenswrapper[4735]: I1122 08:27:15.665144 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerDied","Data":"e733c893683860b8b01e0a8d7c386a05a3e7e3d454f9372451a397057d27eb03"} Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.113216 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.113445 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-log" containerID="cri-o://3ea3e49962d94d8965c1601e2214a5f0749d4f79ae16eae908bbe0250c79ec01" gracePeriod=30 Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.113575 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-api" containerID="cri-o://7188d585575f843c8658f5c4893b2526cbead83edfc8591c27a190f37987d44d" gracePeriod=30 Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.131124 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.131180 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.676996 4735 generic.go:334] "Generic (PLEG): container finished" podID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerID="f85c38b3353bd1397b04cfd53a2500e92a3f2ed267c4f25015dbadbdb86e6110" exitCode=0 Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.677238 4735 generic.go:334] "Generic (PLEG): container finished" podID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerID="c3f1413acf03125c23f14a0669138f9e868c3bb37dd90b6159ae0674994cd5f7" exitCode=0 Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.677068 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerDied","Data":"f85c38b3353bd1397b04cfd53a2500e92a3f2ed267c4f25015dbadbdb86e6110"} Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.677344 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerDied","Data":"c3f1413acf03125c23f14a0669138f9e868c3bb37dd90b6159ae0674994cd5f7"} Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.679177 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" event={"ID":"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb","Type":"ContainerStarted","Data":"9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851"} Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.679331 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.681011 4735 generic.go:334] "Generic (PLEG): container finished" podID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerID="3ea3e49962d94d8965c1601e2214a5f0749d4f79ae16eae908bbe0250c79ec01" exitCode=143 Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.681052 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c53b788-2694-4f84-8ad5-e1f1b17b531d","Type":"ContainerDied","Data":"3ea3e49962d94d8965c1601e2214a5f0749d4f79ae16eae908bbe0250c79ec01"} Nov 22 08:27:16 crc kubenswrapper[4735]: I1122 08:27:16.699973 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" podStartSLOduration=4.69995657 podStartE2EDuration="4.69995657s" podCreationTimestamp="2025-11-22 08:27:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:27:16.699709613 +0000 UTC m=+1458.304048218" watchObservedRunningTime="2025-11-22 08:27:16.69995657 +0000 UTC m=+1458.304295175" Nov 22 08:27:18 crc kubenswrapper[4735]: I1122 08:27:18.572787 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:27:18 crc kubenswrapper[4735]: I1122 08:27:18.635261 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:27:18 crc kubenswrapper[4735]: I1122 08:27:18.826373 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h7slb"] Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.039201 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.061567 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.752303 4735 generic.go:334] "Generic (PLEG): container finished" podID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerID="4e60d4add92bb9b9b00d433eaeef48d6d48a9732b6001d661db7ec53e457daaf" exitCode=0 Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.752366 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerDied","Data":"4e60d4add92bb9b9b00d433eaeef48d6d48a9732b6001d661db7ec53e457daaf"} Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.755086 4735 generic.go:334] "Generic (PLEG): container finished" podID="4179aca1-9f3a-4cd5-a9ad-3a301a313668" containerID="0b072957e6f7373a5320c3e2a92f347bb471db0944c469c4e92d077ccc93a5e0" exitCode=0 Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.755153 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-rkfmf" event={"ID":"4179aca1-9f3a-4cd5-a9ad-3a301a313668","Type":"ContainerDied","Data":"0b072957e6f7373a5320c3e2a92f347bb471db0944c469c4e92d077ccc93a5e0"} Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.757122 4735 generic.go:334] "Generic (PLEG): container finished" podID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerID="7188d585575f843c8658f5c4893b2526cbead83edfc8591c27a190f37987d44d" exitCode=0 Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.757206 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c53b788-2694-4f84-8ad5-e1f1b17b531d","Type":"ContainerDied","Data":"7188d585575f843c8658f5c4893b2526cbead83edfc8591c27a190f37987d44d"} Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.757339 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h7slb" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="registry-server" containerID="cri-o://92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3" gracePeriod=2 Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.777373 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.989586 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-8grl4"] Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.991174 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.994701 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 22 08:27:19 crc kubenswrapper[4735]: I1122 08:27:19.995374 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.018221 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8grl4"] Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.038685 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.045356 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.049135 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-scripts\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.049234 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2kxv\" (UniqueName: \"kubernetes.io/projected/272926fe-e620-4d53-b17f-524b80bdce0d-kube-api-access-z2kxv\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.049401 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-config-data\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.049433 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.151300 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-scripts\") pod \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.151378 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-log-httpd\") pod \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.151415 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z26b\" (UniqueName: \"kubernetes.io/projected/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-kube-api-access-8z26b\") pod \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.151614 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-combined-ca-bundle\") pod \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.151801 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-combined-ca-bundle\") pod \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.151825 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-run-httpd\") pod \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.151861 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-sg-core-conf-yaml\") pod \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.151890 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53b788-2694-4f84-8ad5-e1f1b17b531d-logs\") pod \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.152037 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-config-data\") pod \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.152110 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-config-data\") pod \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\" (UID: \"69fdd40e-0c07-49c4-a65f-e5be20afbf5d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.152135 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c6rk\" (UniqueName: \"kubernetes.io/projected/8c53b788-2694-4f84-8ad5-e1f1b17b531d-kube-api-access-9c6rk\") pod \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\" (UID: \"8c53b788-2694-4f84-8ad5-e1f1b17b531d\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.152740 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-scripts\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.152866 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2kxv\" (UniqueName: \"kubernetes.io/projected/272926fe-e620-4d53-b17f-524b80bdce0d-kube-api-access-z2kxv\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.153039 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-config-data\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.153073 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.153717 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "69fdd40e-0c07-49c4-a65f-e5be20afbf5d" (UID: "69fdd40e-0c07-49c4-a65f-e5be20afbf5d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.162898 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.163160 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-scripts" (OuterVolumeSpecName: "scripts") pod "69fdd40e-0c07-49c4-a65f-e5be20afbf5d" (UID: "69fdd40e-0c07-49c4-a65f-e5be20afbf5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.169544 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c53b788-2694-4f84-8ad5-e1f1b17b531d-kube-api-access-9c6rk" (OuterVolumeSpecName: "kube-api-access-9c6rk") pod "8c53b788-2694-4f84-8ad5-e1f1b17b531d" (UID: "8c53b788-2694-4f84-8ad5-e1f1b17b531d"). InnerVolumeSpecName "kube-api-access-9c6rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.170131 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c53b788-2694-4f84-8ad5-e1f1b17b531d-logs" (OuterVolumeSpecName: "logs") pod "8c53b788-2694-4f84-8ad5-e1f1b17b531d" (UID: "8c53b788-2694-4f84-8ad5-e1f1b17b531d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.170165 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-kube-api-access-8z26b" (OuterVolumeSpecName: "kube-api-access-8z26b") pod "69fdd40e-0c07-49c4-a65f-e5be20afbf5d" (UID: "69fdd40e-0c07-49c4-a65f-e5be20afbf5d"). InnerVolumeSpecName "kube-api-access-8z26b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.170419 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "69fdd40e-0c07-49c4-a65f-e5be20afbf5d" (UID: "69fdd40e-0c07-49c4-a65f-e5be20afbf5d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.184886 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-scripts\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.190612 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2kxv\" (UniqueName: \"kubernetes.io/projected/272926fe-e620-4d53-b17f-524b80bdce0d-kube-api-access-z2kxv\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.191617 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-config-data\") pod \"nova-cell1-cell-mapping-8grl4\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.232718 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "69fdd40e-0c07-49c4-a65f-e5be20afbf5d" (UID: "69fdd40e-0c07-49c4-a65f-e5be20afbf5d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.245664 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-config-data" (OuterVolumeSpecName: "config-data") pod "8c53b788-2694-4f84-8ad5-e1f1b17b531d" (UID: "8c53b788-2694-4f84-8ad5-e1f1b17b531d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.258615 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.258664 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.258678 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z26b\" (UniqueName: \"kubernetes.io/projected/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-kube-api-access-8z26b\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.260416 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.260439 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.260489 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53b788-2694-4f84-8ad5-e1f1b17b531d-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.260503 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.260516 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c6rk\" (UniqueName: \"kubernetes.io/projected/8c53b788-2694-4f84-8ad5-e1f1b17b531d-kube-api-access-9c6rk\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.285793 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c53b788-2694-4f84-8ad5-e1f1b17b531d" (UID: "8c53b788-2694-4f84-8ad5-e1f1b17b531d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.300654 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69fdd40e-0c07-49c4-a65f-e5be20afbf5d" (UID: "69fdd40e-0c07-49c4-a65f-e5be20afbf5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.332323 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.346636 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-config-data" (OuterVolumeSpecName: "config-data") pod "69fdd40e-0c07-49c4-a65f-e5be20afbf5d" (UID: "69fdd40e-0c07-49c4-a65f-e5be20afbf5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.367765 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.369325 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.369356 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69fdd40e-0c07-49c4-a65f-e5be20afbf5d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.369366 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53b788-2694-4f84-8ad5-e1f1b17b531d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.470593 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-catalog-content\") pod \"ca123e9d-79c1-4c31-b633-59c1ba594be7\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.471154 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x68qt\" (UniqueName: \"kubernetes.io/projected/ca123e9d-79c1-4c31-b633-59c1ba594be7-kube-api-access-x68qt\") pod \"ca123e9d-79c1-4c31-b633-59c1ba594be7\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.471192 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-utilities\") pod \"ca123e9d-79c1-4c31-b633-59c1ba594be7\" (UID: \"ca123e9d-79c1-4c31-b633-59c1ba594be7\") " Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.480939 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca123e9d-79c1-4c31-b633-59c1ba594be7-kube-api-access-x68qt" (OuterVolumeSpecName: "kube-api-access-x68qt") pod "ca123e9d-79c1-4c31-b633-59c1ba594be7" (UID: "ca123e9d-79c1-4c31-b633-59c1ba594be7"). InnerVolumeSpecName "kube-api-access-x68qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.497057 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-utilities" (OuterVolumeSpecName: "utilities") pod "ca123e9d-79c1-4c31-b633-59c1ba594be7" (UID: "ca123e9d-79c1-4c31-b633-59c1ba594be7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.534933 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca123e9d-79c1-4c31-b633-59c1ba594be7" (UID: "ca123e9d-79c1-4c31-b633-59c1ba594be7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.574193 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.574450 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x68qt\" (UniqueName: \"kubernetes.io/projected/ca123e9d-79c1-4c31-b633-59c1ba594be7-kube-api-access-x68qt\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.574613 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca123e9d-79c1-4c31-b633-59c1ba594be7-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.774029 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"69fdd40e-0c07-49c4-a65f-e5be20afbf5d","Type":"ContainerDied","Data":"4a9291bd350bf65d69a5d25deae03b8d351f5202c0833faaf932245ed5612360"} Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.774313 4735 scope.go:117] "RemoveContainer" containerID="f85c38b3353bd1397b04cfd53a2500e92a3f2ed267c4f25015dbadbdb86e6110" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.774082 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.777957 4735 generic.go:334] "Generic (PLEG): container finished" podID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerID="92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3" exitCode=0 Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.778014 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7slb" event={"ID":"ca123e9d-79c1-4c31-b633-59c1ba594be7","Type":"ContainerDied","Data":"92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3"} Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.778039 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7slb" event={"ID":"ca123e9d-79c1-4c31-b633-59c1ba594be7","Type":"ContainerDied","Data":"3eff12aa7d5202cd20d8b48cd9fc12c099d34512db5fad1969ecc734cfb3bdec"} Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.778093 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h7slb" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.785350 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.786052 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c53b788-2694-4f84-8ad5-e1f1b17b531d","Type":"ContainerDied","Data":"186f87c04a76d7262577ad0a3a95822ec2be38dd96c39e27de25e74bdcda48e2"} Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.804971 4735 scope.go:117] "RemoveContainer" containerID="e733c893683860b8b01e0a8d7c386a05a3e7e3d454f9372451a397057d27eb03" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.829224 4735 scope.go:117] "RemoveContainer" containerID="4e60d4add92bb9b9b00d433eaeef48d6d48a9732b6001d661db7ec53e457daaf" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.840708 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.852856 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.866830 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.876922 4735 scope.go:117] "RemoveContainer" containerID="c3f1413acf03125c23f14a0669138f9e868c3bb37dd90b6159ae0674994cd5f7" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.883082 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.909403 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h7slb"] Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.926417 4735 scope.go:117] "RemoveContainer" containerID="92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.968858 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.969883 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-log" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.969902 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-log" Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.969955 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-api" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.969963 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-api" Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.969983 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="extract-content" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.969994 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="extract-content" Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.970034 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="sg-core" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.970041 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="sg-core" Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.970084 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="extract-utilities" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.970097 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="extract-utilities" Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.970111 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="ceilometer-central-agent" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.970119 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="ceilometer-central-agent" Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.970139 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="registry-server" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.970147 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="registry-server" Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.970178 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="proxy-httpd" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.970187 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="proxy-httpd" Nov 22 08:27:20 crc kubenswrapper[4735]: E1122 08:27:20.970211 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="ceilometer-notification-agent" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.970218 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="ceilometer-notification-agent" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.980219 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="ceilometer-notification-agent" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.980258 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-log" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.980285 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="proxy-httpd" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.980311 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" containerName="registry-server" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.980333 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" containerName="nova-api-api" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.980355 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="ceilometer-central-agent" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.980363 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="sg-core" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.992064 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:27:20 crc kubenswrapper[4735]: I1122 08:27:20.997689 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h7slb"] Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.000546 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.000796 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.001800 4735 scope.go:117] "RemoveContainer" containerID="29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.044588 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.046940 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.050450 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.050863 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.051197 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.057776 4735 scope.go:117] "RemoveContainer" containerID="90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.094437 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.095879 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgrg2\" (UniqueName: \"kubernetes.io/projected/efc29c3b-bffc-4aac-81ff-d26313da9c78-kube-api-access-cgrg2\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.095925 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrvm4\" (UniqueName: \"kubernetes.io/projected/db97a667-6978-4ac8-9e54-0648ddb62583-kube-api-access-rrvm4\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.095952 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.095979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-config-data\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096006 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-public-tls-certs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096027 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096057 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-log-httpd\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096111 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-config-data\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096139 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-run-httpd\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096169 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc29c3b-bffc-4aac-81ff-d26313da9c78-logs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096188 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-scripts\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096205 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.096231 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-internal-tls-certs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.132765 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.146785 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8grl4"] Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.161781 4735 scope.go:117] "RemoveContainer" containerID="92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3" Nov 22 08:27:21 crc kubenswrapper[4735]: E1122 08:27:21.162157 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3\": container with ID starting with 92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3 not found: ID does not exist" containerID="92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.162190 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3"} err="failed to get container status \"92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3\": rpc error: code = NotFound desc = could not find container \"92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3\": container with ID starting with 92aed8f70078c26dabedd2ececa182c21df42ab2a9a339bd52d5ab42c1c1b0a3 not found: ID does not exist" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.162210 4735 scope.go:117] "RemoveContainer" containerID="29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc" Nov 22 08:27:21 crc kubenswrapper[4735]: E1122 08:27:21.162487 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc\": container with ID starting with 29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc not found: ID does not exist" containerID="29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.162506 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc"} err="failed to get container status \"29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc\": rpc error: code = NotFound desc = could not find container \"29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc\": container with ID starting with 29561532ae92955f4cbe7ccf1296cb651d63a1e21625fc5a017d476a9d3b5bdc not found: ID does not exist" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.162517 4735 scope.go:117] "RemoveContainer" containerID="90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697" Nov 22 08:27:21 crc kubenswrapper[4735]: E1122 08:27:21.162694 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697\": container with ID starting with 90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697 not found: ID does not exist" containerID="90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.162716 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697"} err="failed to get container status \"90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697\": rpc error: code = NotFound desc = could not find container \"90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697\": container with ID starting with 90bff2628aa7349de60a65063f03378bd1ff27fbfc81bbdaeb2cb0ccad4de697 not found: ID does not exist" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.162738 4735 scope.go:117] "RemoveContainer" containerID="7188d585575f843c8658f5c4893b2526cbead83edfc8591c27a190f37987d44d" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199022 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgrg2\" (UniqueName: \"kubernetes.io/projected/efc29c3b-bffc-4aac-81ff-d26313da9c78-kube-api-access-cgrg2\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199079 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrvm4\" (UniqueName: \"kubernetes.io/projected/db97a667-6978-4ac8-9e54-0648ddb62583-kube-api-access-rrvm4\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199111 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199141 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-config-data\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199181 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-public-tls-certs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199211 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199256 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-log-httpd\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199348 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-config-data\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199658 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-run-httpd\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199710 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc29c3b-bffc-4aac-81ff-d26313da9c78-logs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199733 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-scripts\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.201054 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.201138 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-internal-tls-certs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.201394 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-run-httpd\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.199930 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-log-httpd\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.204898 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-scripts\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.206683 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc29c3b-bffc-4aac-81ff-d26313da9c78-logs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.206688 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.207584 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-config-data\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.208437 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-config-data\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.216217 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.216718 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.217049 4735 scope.go:117] "RemoveContainer" containerID="3ea3e49962d94d8965c1601e2214a5f0749d4f79ae16eae908bbe0250c79ec01" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.217362 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgrg2\" (UniqueName: \"kubernetes.io/projected/efc29c3b-bffc-4aac-81ff-d26313da9c78-kube-api-access-cgrg2\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.217489 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-public-tls-certs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.219342 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrvm4\" (UniqueName: \"kubernetes.io/projected/db97a667-6978-4ac8-9e54-0648ddb62583-kube-api-access-rrvm4\") pod \"ceilometer-0\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.228191 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-internal-tls-certs\") pod \"nova-api-0\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.240731 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.286752 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" path="/var/lib/kubelet/pods/69fdd40e-0c07-49c4-a65f-e5be20afbf5d/volumes" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.287775 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c53b788-2694-4f84-8ad5-e1f1b17b531d" path="/var/lib/kubelet/pods/8c53b788-2694-4f84-8ad5-e1f1b17b531d/volumes" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.288399 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca123e9d-79c1-4c31-b633-59c1ba594be7" path="/var/lib/kubelet/pods/ca123e9d-79c1-4c31-b633-59c1ba594be7/volumes" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.302793 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-combined-ca-bundle\") pod \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.302849 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdztc\" (UniqueName: \"kubernetes.io/projected/4179aca1-9f3a-4cd5-a9ad-3a301a313668-kube-api-access-xdztc\") pod \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.302882 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-config-data\") pod \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.303025 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-scripts\") pod \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\" (UID: \"4179aca1-9f3a-4cd5-a9ad-3a301a313668\") " Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.312373 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-scripts" (OuterVolumeSpecName: "scripts") pod "4179aca1-9f3a-4cd5-a9ad-3a301a313668" (UID: "4179aca1-9f3a-4cd5-a9ad-3a301a313668"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.312924 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4179aca1-9f3a-4cd5-a9ad-3a301a313668-kube-api-access-xdztc" (OuterVolumeSpecName: "kube-api-access-xdztc") pod "4179aca1-9f3a-4cd5-a9ad-3a301a313668" (UID: "4179aca1-9f3a-4cd5-a9ad-3a301a313668"). InnerVolumeSpecName "kube-api-access-xdztc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.356156 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-config-data" (OuterVolumeSpecName: "config-data") pod "4179aca1-9f3a-4cd5-a9ad-3a301a313668" (UID: "4179aca1-9f3a-4cd5-a9ad-3a301a313668"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.366092 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4179aca1-9f3a-4cd5-a9ad-3a301a313668" (UID: "4179aca1-9f3a-4cd5-a9ad-3a301a313668"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.404131 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.405803 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.405842 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdztc\" (UniqueName: \"kubernetes.io/projected/4179aca1-9f3a-4cd5-a9ad-3a301a313668-kube-api-access-xdztc\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.405854 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.405862 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4179aca1-9f3a-4cd5-a9ad-3a301a313668-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.441232 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.441861 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.805377 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8grl4" event={"ID":"272926fe-e620-4d53-b17f-524b80bdce0d","Type":"ContainerStarted","Data":"ae213ed10166fc9ce8fa46d4d6004f5952b5097ca6c62452515bfe43001b5df8"} Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.805963 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8grl4" event={"ID":"272926fe-e620-4d53-b17f-524b80bdce0d","Type":"ContainerStarted","Data":"b836525c49c523452ca9c2efccdb6b99ab9f99913e4a34c68722f5f5342982d4"} Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.813230 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-rkfmf" event={"ID":"4179aca1-9f3a-4cd5-a9ad-3a301a313668","Type":"ContainerDied","Data":"09f29b8972850cdd0b6d19e4174fae1c64ed70a977f6a4513a01fbbec135e502"} Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.813261 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-rkfmf" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.813270 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09f29b8972850cdd0b6d19e4174fae1c64ed70a977f6a4513a01fbbec135e502" Nov 22 08:27:21 crc kubenswrapper[4735]: I1122 08:27:21.838958 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-8grl4" podStartSLOduration=2.838938383 podStartE2EDuration="2.838938383s" podCreationTimestamp="2025-11-22 08:27:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:27:21.825322241 +0000 UTC m=+1463.429660846" watchObservedRunningTime="2025-11-22 08:27:21.838938383 +0000 UTC m=+1463.443276988" Nov 22 08:27:22 crc kubenswrapper[4735]: I1122 08:27:22.039632 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:22 crc kubenswrapper[4735]: I1122 08:27:22.116473 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:22 crc kubenswrapper[4735]: W1122 08:27:22.128788 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb97a667_6978_4ac8_9e54_0648ddb62583.slice/crio-78ee6de4271b91e5781dcf3cedab1e17ea5124c699fb3090f64063ae2740ccca WatchSource:0}: Error finding container 78ee6de4271b91e5781dcf3cedab1e17ea5124c699fb3090f64063ae2740ccca: Status 404 returned error can't find the container with id 78ee6de4271b91e5781dcf3cedab1e17ea5124c699fb3090f64063ae2740ccca Nov 22 08:27:22 crc kubenswrapper[4735]: I1122 08:27:22.835290 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerStarted","Data":"78ee6de4271b91e5781dcf3cedab1e17ea5124c699fb3090f64063ae2740ccca"} Nov 22 08:27:22 crc kubenswrapper[4735]: I1122 08:27:22.839895 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc29c3b-bffc-4aac-81ff-d26313da9c78","Type":"ContainerStarted","Data":"698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a"} Nov 22 08:27:22 crc kubenswrapper[4735]: I1122 08:27:22.839940 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc29c3b-bffc-4aac-81ff-d26313da9c78","Type":"ContainerStarted","Data":"7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9"} Nov 22 08:27:22 crc kubenswrapper[4735]: I1122 08:27:22.839949 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc29c3b-bffc-4aac-81ff-d26313da9c78","Type":"ContainerStarted","Data":"9bc6da45354599733789f71ea347b9d17e5bc7d14a22fac81b9dcd4ae1c733bf"} Nov 22 08:27:22 crc kubenswrapper[4735]: I1122 08:27:22.874542 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.874521048 podStartE2EDuration="2.874521048s" podCreationTimestamp="2025-11-22 08:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:27:22.864551145 +0000 UTC m=+1464.468889770" watchObservedRunningTime="2025-11-22 08:27:22.874521048 +0000 UTC m=+1464.478859653" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.117631 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.203916 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6p8hd"] Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.204148 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" podUID="58815e92-d627-4845-99df-143d68ba6bd2" containerName="dnsmasq-dns" containerID="cri-o://9ffbeb65bba7d6a85c4e62e2a114e21faf6505ccb5783a05f61cf8230dc60f47" gracePeriod=10 Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.333339 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 22 08:27:23 crc kubenswrapper[4735]: E1122 08:27:23.334471 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4179aca1-9f3a-4cd5-a9ad-3a301a313668" containerName="aodh-db-sync" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.334494 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4179aca1-9f3a-4cd5-a9ad-3a301a313668" containerName="aodh-db-sync" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.335112 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4179aca1-9f3a-4cd5-a9ad-3a301a313668" containerName="aodh-db-sync" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.357989 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.358226 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.377098 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.377454 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.378285 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-bfgs7" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.386851 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" podUID="58815e92-d627-4845-99df-143d68ba6bd2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.236:5353: connect: connection refused" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.482938 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-config-data\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.483044 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-scripts\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.483122 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-combined-ca-bundle\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.483144 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjq6t\" (UniqueName: \"kubernetes.io/projected/477b15d5-59f1-4f08-97eb-14ae37feb4ae-kube-api-access-fjq6t\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.585177 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-combined-ca-bundle\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.585439 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjq6t\" (UniqueName: \"kubernetes.io/projected/477b15d5-59f1-4f08-97eb-14ae37feb4ae-kube-api-access-fjq6t\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.585635 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-config-data\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.585714 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-scripts\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.611891 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-config-data\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.612212 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-scripts\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.618078 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-combined-ca-bundle\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.633080 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjq6t\" (UniqueName: \"kubernetes.io/projected/477b15d5-59f1-4f08-97eb-14ae37feb4ae-kube-api-access-fjq6t\") pod \"aodh-0\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.742066 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.868926 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerStarted","Data":"cbe8a7d938571463e9a3286fbf70433fcb477eb1bbf0da8d194b140f6d576a46"} Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.874820 4735 generic.go:334] "Generic (PLEG): container finished" podID="58815e92-d627-4845-99df-143d68ba6bd2" containerID="9ffbeb65bba7d6a85c4e62e2a114e21faf6505ccb5783a05f61cf8230dc60f47" exitCode=0 Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.875807 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" event={"ID":"58815e92-d627-4845-99df-143d68ba6bd2","Type":"ContainerDied","Data":"9ffbeb65bba7d6a85c4e62e2a114e21faf6505ccb5783a05f61cf8230dc60f47"} Nov 22 08:27:23 crc kubenswrapper[4735]: I1122 08:27:23.953158 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.102255 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-svc\") pod \"58815e92-d627-4845-99df-143d68ba6bd2\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.102586 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-sb\") pod \"58815e92-d627-4845-99df-143d68ba6bd2\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.102671 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-swift-storage-0\") pod \"58815e92-d627-4845-99df-143d68ba6bd2\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.102759 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-config\") pod \"58815e92-d627-4845-99df-143d68ba6bd2\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.102788 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dtkc\" (UniqueName: \"kubernetes.io/projected/58815e92-d627-4845-99df-143d68ba6bd2-kube-api-access-4dtkc\") pod \"58815e92-d627-4845-99df-143d68ba6bd2\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.102917 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-nb\") pod \"58815e92-d627-4845-99df-143d68ba6bd2\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.115112 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58815e92-d627-4845-99df-143d68ba6bd2-kube-api-access-4dtkc" (OuterVolumeSpecName: "kube-api-access-4dtkc") pod "58815e92-d627-4845-99df-143d68ba6bd2" (UID: "58815e92-d627-4845-99df-143d68ba6bd2"). InnerVolumeSpecName "kube-api-access-4dtkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.212187 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "58815e92-d627-4845-99df-143d68ba6bd2" (UID: "58815e92-d627-4845-99df-143d68ba6bd2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.212804 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-swift-storage-0\") pod \"58815e92-d627-4845-99df-143d68ba6bd2\" (UID: \"58815e92-d627-4845-99df-143d68ba6bd2\") " Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.214063 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dtkc\" (UniqueName: \"kubernetes.io/projected/58815e92-d627-4845-99df-143d68ba6bd2-kube-api-access-4dtkc\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:24 crc kubenswrapper[4735]: W1122 08:27:24.212918 4735 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/58815e92-d627-4845-99df-143d68ba6bd2/volumes/kubernetes.io~configmap/dns-swift-storage-0 Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.214278 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "58815e92-d627-4845-99df-143d68ba6bd2" (UID: "58815e92-d627-4845-99df-143d68ba6bd2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.215085 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "58815e92-d627-4845-99df-143d68ba6bd2" (UID: "58815e92-d627-4845-99df-143d68ba6bd2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.240188 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "58815e92-d627-4845-99df-143d68ba6bd2" (UID: "58815e92-d627-4845-99df-143d68ba6bd2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.262733 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "58815e92-d627-4845-99df-143d68ba6bd2" (UID: "58815e92-d627-4845-99df-143d68ba6bd2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.278108 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-config" (OuterVolumeSpecName: "config") pod "58815e92-d627-4845-99df-143d68ba6bd2" (UID: "58815e92-d627-4845-99df-143d68ba6bd2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.317560 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.317608 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.317621 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.317633 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.317645 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58815e92-d627-4845-99df-143d68ba6bd2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.440249 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.907384 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerStarted","Data":"62182e2ac75e8783c7308cb955c32951edfbd0ee45431198ef15199de82f2915"} Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.909552 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerStarted","Data":"62c7ed5cc255c2d13288c6e993caa464a525d2e5f44653c8e3201fa895fd92b6"} Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.909577 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerStarted","Data":"e5c3ce8eaa71fa1edeca6bc7b9feffa720746a2de6191b8d8c657ff949f0ca73"} Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.928124 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" event={"ID":"58815e92-d627-4845-99df-143d68ba6bd2","Type":"ContainerDied","Data":"140ec08fc8597756902cfe29833e4e6d4cd4c1888927ee1b701341105bd8b3b9"} Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.928169 4735 scope.go:117] "RemoveContainer" containerID="9ffbeb65bba7d6a85c4e62e2a114e21faf6505ccb5783a05f61cf8230dc60f47" Nov 22 08:27:24 crc kubenswrapper[4735]: I1122 08:27:24.928331 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-6p8hd" Nov 22 08:27:25 crc kubenswrapper[4735]: I1122 08:27:25.035534 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6p8hd"] Nov 22 08:27:25 crc kubenswrapper[4735]: I1122 08:27:25.044718 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6p8hd"] Nov 22 08:27:25 crc kubenswrapper[4735]: I1122 08:27:25.137274 4735 scope.go:117] "RemoveContainer" containerID="29db1d9c7a41cc25f47cf5168ada56d7455374ae0964ca0429b4f116e0deba1e" Nov 22 08:27:25 crc kubenswrapper[4735]: I1122 08:27:25.283264 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58815e92-d627-4845-99df-143d68ba6bd2" path="/var/lib/kubelet/pods/58815e92-d627-4845-99df-143d68ba6bd2/volumes" Nov 22 08:27:25 crc kubenswrapper[4735]: I1122 08:27:25.942795 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerStarted","Data":"13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d"} Nov 22 08:27:26 crc kubenswrapper[4735]: I1122 08:27:26.964354 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerStarted","Data":"91167e8c5a2e16ea1f877d5eb7b52bbe421ad3cc8a8978f172f9eaa12dea1c87"} Nov 22 08:27:26 crc kubenswrapper[4735]: I1122 08:27:26.964768 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="ceilometer-central-agent" containerID="cri-o://cbe8a7d938571463e9a3286fbf70433fcb477eb1bbf0da8d194b140f6d576a46" gracePeriod=30 Nov 22 08:27:26 crc kubenswrapper[4735]: I1122 08:27:26.964808 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="ceilometer-notification-agent" containerID="cri-o://e5c3ce8eaa71fa1edeca6bc7b9feffa720746a2de6191b8d8c657ff949f0ca73" gracePeriod=30 Nov 22 08:27:26 crc kubenswrapper[4735]: I1122 08:27:26.964846 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="proxy-httpd" containerID="cri-o://91167e8c5a2e16ea1f877d5eb7b52bbe421ad3cc8a8978f172f9eaa12dea1c87" gracePeriod=30 Nov 22 08:27:26 crc kubenswrapper[4735]: I1122 08:27:26.964882 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="sg-core" containerID="cri-o://62c7ed5cc255c2d13288c6e993caa464a525d2e5f44653c8e3201fa895fd92b6" gracePeriod=30 Nov 22 08:27:26 crc kubenswrapper[4735]: I1122 08:27:26.966533 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:27:26 crc kubenswrapper[4735]: I1122 08:27:26.992862 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.450178703 podStartE2EDuration="6.992838243s" podCreationTimestamp="2025-11-22 08:27:20 +0000 UTC" firstStartedPulling="2025-11-22 08:27:22.146051713 +0000 UTC m=+1463.750390318" lastFinishedPulling="2025-11-22 08:27:25.688711253 +0000 UTC m=+1467.293049858" observedRunningTime="2025-11-22 08:27:26.990390236 +0000 UTC m=+1468.594728871" watchObservedRunningTime="2025-11-22 08:27:26.992838243 +0000 UTC m=+1468.597176868" Nov 22 08:27:27 crc kubenswrapper[4735]: I1122 08:27:27.977215 4735 generic.go:334] "Generic (PLEG): container finished" podID="db97a667-6978-4ac8-9e54-0648ddb62583" containerID="91167e8c5a2e16ea1f877d5eb7b52bbe421ad3cc8a8978f172f9eaa12dea1c87" exitCode=0 Nov 22 08:27:27 crc kubenswrapper[4735]: I1122 08:27:27.977533 4735 generic.go:334] "Generic (PLEG): container finished" podID="db97a667-6978-4ac8-9e54-0648ddb62583" containerID="62c7ed5cc255c2d13288c6e993caa464a525d2e5f44653c8e3201fa895fd92b6" exitCode=2 Nov 22 08:27:27 crc kubenswrapper[4735]: I1122 08:27:27.977549 4735 generic.go:334] "Generic (PLEG): container finished" podID="db97a667-6978-4ac8-9e54-0648ddb62583" containerID="e5c3ce8eaa71fa1edeca6bc7b9feffa720746a2de6191b8d8c657ff949f0ca73" exitCode=0 Nov 22 08:27:27 crc kubenswrapper[4735]: I1122 08:27:27.977589 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerDied","Data":"91167e8c5a2e16ea1f877d5eb7b52bbe421ad3cc8a8978f172f9eaa12dea1c87"} Nov 22 08:27:27 crc kubenswrapper[4735]: I1122 08:27:27.977617 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerDied","Data":"62c7ed5cc255c2d13288c6e993caa464a525d2e5f44653c8e3201fa895fd92b6"} Nov 22 08:27:27 crc kubenswrapper[4735]: I1122 08:27:27.977628 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerDied","Data":"e5c3ce8eaa71fa1edeca6bc7b9feffa720746a2de6191b8d8c657ff949f0ca73"} Nov 22 08:27:28 crc kubenswrapper[4735]: I1122 08:27:28.002822 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 22 08:27:28 crc kubenswrapper[4735]: I1122 08:27:28.989197 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerStarted","Data":"90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e"} Nov 22 08:27:30 crc kubenswrapper[4735]: I1122 08:27:30.006240 4735 generic.go:334] "Generic (PLEG): container finished" podID="272926fe-e620-4d53-b17f-524b80bdce0d" containerID="ae213ed10166fc9ce8fa46d4d6004f5952b5097ca6c62452515bfe43001b5df8" exitCode=0 Nov 22 08:27:30 crc kubenswrapper[4735]: I1122 08:27:30.006283 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8grl4" event={"ID":"272926fe-e620-4d53-b17f-524b80bdce0d","Type":"ContainerDied","Data":"ae213ed10166fc9ce8fa46d4d6004f5952b5097ca6c62452515bfe43001b5df8"} Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.022561 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerStarted","Data":"476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6"} Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.464143 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.464462 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.639876 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.770861 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-combined-ca-bundle\") pod \"272926fe-e620-4d53-b17f-524b80bdce0d\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.771014 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-config-data\") pod \"272926fe-e620-4d53-b17f-524b80bdce0d\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.771058 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2kxv\" (UniqueName: \"kubernetes.io/projected/272926fe-e620-4d53-b17f-524b80bdce0d-kube-api-access-z2kxv\") pod \"272926fe-e620-4d53-b17f-524b80bdce0d\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.771208 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-scripts\") pod \"272926fe-e620-4d53-b17f-524b80bdce0d\" (UID: \"272926fe-e620-4d53-b17f-524b80bdce0d\") " Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.778605 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-scripts" (OuterVolumeSpecName: "scripts") pod "272926fe-e620-4d53-b17f-524b80bdce0d" (UID: "272926fe-e620-4d53-b17f-524b80bdce0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.807653 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/272926fe-e620-4d53-b17f-524b80bdce0d-kube-api-access-z2kxv" (OuterVolumeSpecName: "kube-api-access-z2kxv") pod "272926fe-e620-4d53-b17f-524b80bdce0d" (UID: "272926fe-e620-4d53-b17f-524b80bdce0d"). InnerVolumeSpecName "kube-api-access-z2kxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.815681 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "272926fe-e620-4d53-b17f-524b80bdce0d" (UID: "272926fe-e620-4d53-b17f-524b80bdce0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.825420 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-config-data" (OuterVolumeSpecName: "config-data") pod "272926fe-e620-4d53-b17f-524b80bdce0d" (UID: "272926fe-e620-4d53-b17f-524b80bdce0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.873648 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.873695 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.873712 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272926fe-e620-4d53-b17f-524b80bdce0d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:31 crc kubenswrapper[4735]: I1122 08:27:31.873728 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2kxv\" (UniqueName: \"kubernetes.io/projected/272926fe-e620-4d53-b17f-524b80bdce0d-kube-api-access-z2kxv\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.035588 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8grl4" event={"ID":"272926fe-e620-4d53-b17f-524b80bdce0d","Type":"ContainerDied","Data":"b836525c49c523452ca9c2efccdb6b99ab9f99913e4a34c68722f5f5342982d4"} Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.035637 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b836525c49c523452ca9c2efccdb6b99ab9f99913e4a34c68722f5f5342982d4" Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.035696 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8grl4" Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.166536 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.166794 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-log" containerID="cri-o://7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9" gracePeriod=30 Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.167211 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-api" containerID="cri-o://698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a" gracePeriod=30 Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.188820 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.189038 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" containerName="nova-scheduler-scheduler" containerID="cri-o://98a9e0b2236cc1e46be7cdc73cf7a000baf80113d8acec7d59ec71d8bb4b5acf" gracePeriod=30 Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.215077 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.215342 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-log" containerID="cri-o://7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7" gracePeriod=30 Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.216600 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-metadata" containerID="cri-o://65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798" gracePeriod=30 Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.228635 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.254:8774/\": EOF" Nov 22 08:27:32 crc kubenswrapper[4735]: I1122 08:27:32.228635 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.254:8774/\": EOF" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.057244 4735 generic.go:334] "Generic (PLEG): container finished" podID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerID="7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9" exitCode=143 Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.057325 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc29c3b-bffc-4aac-81ff-d26313da9c78","Type":"ContainerDied","Data":"7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9"} Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.067427 4735 generic.go:334] "Generic (PLEG): container finished" podID="db97a667-6978-4ac8-9e54-0648ddb62583" containerID="cbe8a7d938571463e9a3286fbf70433fcb477eb1bbf0da8d194b140f6d576a46" exitCode=0 Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.067525 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerDied","Data":"cbe8a7d938571463e9a3286fbf70433fcb477eb1bbf0da8d194b140f6d576a46"} Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.071273 4735 generic.go:334] "Generic (PLEG): container finished" podID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerID="7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7" exitCode=143 Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.071323 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"049fb0cb-f818-4234-a5ee-3bbfb518aec8","Type":"ContainerDied","Data":"7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7"} Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.381923 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.417252 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-run-httpd\") pod \"db97a667-6978-4ac8-9e54-0648ddb62583\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.417395 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-combined-ca-bundle\") pod \"db97a667-6978-4ac8-9e54-0648ddb62583\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.417419 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-log-httpd\") pod \"db97a667-6978-4ac8-9e54-0648ddb62583\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.419091 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "db97a667-6978-4ac8-9e54-0648ddb62583" (UID: "db97a667-6978-4ac8-9e54-0648ddb62583"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.421579 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "db97a667-6978-4ac8-9e54-0648ddb62583" (UID: "db97a667-6978-4ac8-9e54-0648ddb62583"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.520537 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-config-data\") pod \"db97a667-6978-4ac8-9e54-0648ddb62583\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.520648 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-sg-core-conf-yaml\") pod \"db97a667-6978-4ac8-9e54-0648ddb62583\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.520678 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrvm4\" (UniqueName: \"kubernetes.io/projected/db97a667-6978-4ac8-9e54-0648ddb62583-kube-api-access-rrvm4\") pod \"db97a667-6978-4ac8-9e54-0648ddb62583\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.520766 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-scripts\") pod \"db97a667-6978-4ac8-9e54-0648ddb62583\" (UID: \"db97a667-6978-4ac8-9e54-0648ddb62583\") " Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.521943 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.521962 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db97a667-6978-4ac8-9e54-0648ddb62583-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.526284 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-scripts" (OuterVolumeSpecName: "scripts") pod "db97a667-6978-4ac8-9e54-0648ddb62583" (UID: "db97a667-6978-4ac8-9e54-0648ddb62583"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.528833 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db97a667-6978-4ac8-9e54-0648ddb62583-kube-api-access-rrvm4" (OuterVolumeSpecName: "kube-api-access-rrvm4") pod "db97a667-6978-4ac8-9e54-0648ddb62583" (UID: "db97a667-6978-4ac8-9e54-0648ddb62583"). InnerVolumeSpecName "kube-api-access-rrvm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.531660 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db97a667-6978-4ac8-9e54-0648ddb62583" (UID: "db97a667-6978-4ac8-9e54-0648ddb62583"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.572620 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "db97a667-6978-4ac8-9e54-0648ddb62583" (UID: "db97a667-6978-4ac8-9e54-0648ddb62583"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:33 crc kubenswrapper[4735]: E1122 08:27:33.575930 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="98a9e0b2236cc1e46be7cdc73cf7a000baf80113d8acec7d59ec71d8bb4b5acf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 22 08:27:33 crc kubenswrapper[4735]: E1122 08:27:33.577973 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="98a9e0b2236cc1e46be7cdc73cf7a000baf80113d8acec7d59ec71d8bb4b5acf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 22 08:27:33 crc kubenswrapper[4735]: E1122 08:27:33.580150 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="98a9e0b2236cc1e46be7cdc73cf7a000baf80113d8acec7d59ec71d8bb4b5acf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 22 08:27:33 crc kubenswrapper[4735]: E1122 08:27:33.580226 4735 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" containerName="nova-scheduler-scheduler" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.625910 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.625958 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.625970 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrvm4\" (UniqueName: \"kubernetes.io/projected/db97a667-6978-4ac8-9e54-0648ddb62583-kube-api-access-rrvm4\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.625983 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.706291 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-config-data" (OuterVolumeSpecName: "config-data") pod "db97a667-6978-4ac8-9e54-0648ddb62583" (UID: "db97a667-6978-4ac8-9e54-0648ddb62583"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:33 crc kubenswrapper[4735]: I1122 08:27:33.728341 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db97a667-6978-4ac8-9e54-0648ddb62583-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.121202 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerStarted","Data":"af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a"} Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.121527 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-api" containerID="cri-o://13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d" gracePeriod=30 Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.121768 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-notifier" containerID="cri-o://476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6" gracePeriod=30 Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.121952 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-listener" containerID="cri-o://af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a" gracePeriod=30 Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.122073 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-evaluator" containerID="cri-o://90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e" gracePeriod=30 Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.151084 4735 generic.go:334] "Generic (PLEG): container finished" podID="b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" containerID="98a9e0b2236cc1e46be7cdc73cf7a000baf80113d8acec7d59ec71d8bb4b5acf" exitCode=0 Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.151469 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35","Type":"ContainerDied","Data":"98a9e0b2236cc1e46be7cdc73cf7a000baf80113d8acec7d59ec71d8bb4b5acf"} Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.179728 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.827677333 podStartE2EDuration="11.179704189s" podCreationTimestamp="2025-11-22 08:27:23 +0000 UTC" firstStartedPulling="2025-11-22 08:27:24.457873286 +0000 UTC m=+1466.062211891" lastFinishedPulling="2025-11-22 08:27:32.809900142 +0000 UTC m=+1474.414238747" observedRunningTime="2025-11-22 08:27:34.150683315 +0000 UTC m=+1475.755021930" watchObservedRunningTime="2025-11-22 08:27:34.179704189 +0000 UTC m=+1475.784042794" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.186057 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db97a667-6978-4ac8-9e54-0648ddb62583","Type":"ContainerDied","Data":"78ee6de4271b91e5781dcf3cedab1e17ea5124c699fb3090f64063ae2740ccca"} Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.186130 4735 scope.go:117] "RemoveContainer" containerID="91167e8c5a2e16ea1f877d5eb7b52bbe421ad3cc8a8978f172f9eaa12dea1c87" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.186234 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.232873 4735 scope.go:117] "RemoveContainer" containerID="62c7ed5cc255c2d13288c6e993caa464a525d2e5f44653c8e3201fa895fd92b6" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.258574 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.281167 4735 scope.go:117] "RemoveContainer" containerID="e5c3ce8eaa71fa1edeca6bc7b9feffa720746a2de6191b8d8c657ff949f0ca73" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.298663 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.312579 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:34 crc kubenswrapper[4735]: E1122 08:27:34.313198 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="ceilometer-central-agent" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313223 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="ceilometer-central-agent" Nov 22 08:27:34 crc kubenswrapper[4735]: E1122 08:27:34.313238 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="proxy-httpd" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313249 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="proxy-httpd" Nov 22 08:27:34 crc kubenswrapper[4735]: E1122 08:27:34.313278 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="272926fe-e620-4d53-b17f-524b80bdce0d" containerName="nova-manage" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313285 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="272926fe-e620-4d53-b17f-524b80bdce0d" containerName="nova-manage" Nov 22 08:27:34 crc kubenswrapper[4735]: E1122 08:27:34.313303 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58815e92-d627-4845-99df-143d68ba6bd2" containerName="dnsmasq-dns" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313311 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="58815e92-d627-4845-99df-143d68ba6bd2" containerName="dnsmasq-dns" Nov 22 08:27:34 crc kubenswrapper[4735]: E1122 08:27:34.313339 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58815e92-d627-4845-99df-143d68ba6bd2" containerName="init" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313347 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="58815e92-d627-4845-99df-143d68ba6bd2" containerName="init" Nov 22 08:27:34 crc kubenswrapper[4735]: E1122 08:27:34.313373 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="sg-core" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313381 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="sg-core" Nov 22 08:27:34 crc kubenswrapper[4735]: E1122 08:27:34.313391 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="ceilometer-notification-agent" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313398 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="ceilometer-notification-agent" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313670 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="ceilometer-notification-agent" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313704 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="sg-core" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313736 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="proxy-httpd" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313757 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" containerName="ceilometer-central-agent" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313778 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="58815e92-d627-4845-99df-143d68ba6bd2" containerName="dnsmasq-dns" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.313794 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="272926fe-e620-4d53-b17f-524b80bdce0d" containerName="nova-manage" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.330335 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.334954 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.335790 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.338948 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.352668 4735 scope.go:117] "RemoveContainer" containerID="cbe8a7d938571463e9a3286fbf70433fcb477eb1bbf0da8d194b140f6d576a46" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.459809 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llnn9\" (UniqueName: \"kubernetes.io/projected/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-kube-api-access-llnn9\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.459869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-config-data\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.459921 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.459953 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.460077 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-scripts\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.460152 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-run-httpd\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.460174 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-log-httpd\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.564971 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-scripts\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.565052 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-run-httpd\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.565072 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-log-httpd\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.565165 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llnn9\" (UniqueName: \"kubernetes.io/projected/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-kube-api-access-llnn9\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.565210 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-config-data\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.565255 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.565283 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.566987 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-log-httpd\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.569645 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-run-httpd\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.577325 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.579539 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-config-data\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.601421 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llnn9\" (UniqueName: \"kubernetes.io/projected/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-kube-api-access-llnn9\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.603151 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.607827 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-scripts\") pod \"ceilometer-0\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.666345 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.729419 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.754635 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-72gww"] Nov 22 08:27:34 crc kubenswrapper[4735]: E1122 08:27:34.755232 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" containerName="nova-scheduler-scheduler" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.755256 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" containerName="nova-scheduler-scheduler" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.755507 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" containerName="nova-scheduler-scheduler" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.757088 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.827568 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-72gww"] Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.873983 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-combined-ca-bundle\") pod \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.874125 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pv4p\" (UniqueName: \"kubernetes.io/projected/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-kube-api-access-2pv4p\") pod \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.874369 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-config-data\") pod \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\" (UID: \"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35\") " Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.874834 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxlgp\" (UniqueName: \"kubernetes.io/projected/0365659b-65fd-4e91-a5bc-e49007a221b2-kube-api-access-bxlgp\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.874945 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-catalog-content\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.875022 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-utilities\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.904746 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-kube-api-access-2pv4p" (OuterVolumeSpecName: "kube-api-access-2pv4p") pod "b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" (UID: "b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35"). InnerVolumeSpecName "kube-api-access-2pv4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.910432 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-config-data" (OuterVolumeSpecName: "config-data") pod "b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" (UID: "b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.942531 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" (UID: "b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.977240 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxlgp\" (UniqueName: \"kubernetes.io/projected/0365659b-65fd-4e91-a5bc-e49007a221b2-kube-api-access-bxlgp\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.977898 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-catalog-content\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.977971 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-utilities\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.978092 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.978104 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.978121 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pv4p\" (UniqueName: \"kubernetes.io/projected/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35-kube-api-access-2pv4p\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.978686 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-utilities\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:34 crc kubenswrapper[4735]: I1122 08:27:34.978684 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-catalog-content\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.016550 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxlgp\" (UniqueName: \"kubernetes.io/projected/0365659b-65fd-4e91-a5bc-e49007a221b2-kube-api-access-bxlgp\") pod \"redhat-marketplace-72gww\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.094918 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.248281 4735 generic.go:334] "Generic (PLEG): container finished" podID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerID="90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e" exitCode=0 Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.248312 4735 generic.go:334] "Generic (PLEG): container finished" podID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerID="13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d" exitCode=0 Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.248374 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerDied","Data":"90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e"} Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.248401 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerDied","Data":"13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d"} Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.255342 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35","Type":"ContainerDied","Data":"bc3dd1d60faae1185885083a29c196c9cb3c6ada8fff017d2e566e301f2ce115"} Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.255395 4735 scope.go:117] "RemoveContainer" containerID="98a9e0b2236cc1e46be7cdc73cf7a000baf80113d8acec7d59ec71d8bb4b5acf" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.255523 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.293660 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db97a667-6978-4ac8-9e54-0648ddb62583" path="/var/lib/kubelet/pods/db97a667-6978-4ac8-9e54-0648ddb62583/volumes" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.330596 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.352731 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.378201 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.379731 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.383476 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.399862 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:27:35 crc kubenswrapper[4735]: W1122 08:27:35.480294 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89192c39_d1e5_4ca8_a1f2_b4c976ae7802.slice/crio-2f363870df73f338b0337be8bbae9ce32d3c2bcfca834be0ab44e8d9121bea64 WatchSource:0}: Error finding container 2f363870df73f338b0337be8bbae9ce32d3c2bcfca834be0ab44e8d9121bea64: Status 404 returned error can't find the container with id 2f363870df73f338b0337be8bbae9ce32d3c2bcfca834be0ab44e8d9121bea64 Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.481942 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.491590 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq7mc\" (UniqueName: \"kubernetes.io/projected/a47abba7-dce0-433f-a78e-1b3224e07835-kube-api-access-kq7mc\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.491908 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a47abba7-dce0-433f-a78e-1b3224e07835-config-data\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.491981 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a47abba7-dce0-433f-a78e-1b3224e07835-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.595246 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a47abba7-dce0-433f-a78e-1b3224e07835-config-data\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.595350 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a47abba7-dce0-433f-a78e-1b3224e07835-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.595492 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq7mc\" (UniqueName: \"kubernetes.io/projected/a47abba7-dce0-433f-a78e-1b3224e07835-kube-api-access-kq7mc\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.613479 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a47abba7-dce0-433f-a78e-1b3224e07835-config-data\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.613696 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a47abba7-dce0-433f-a78e-1b3224e07835-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.618595 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq7mc\" (UniqueName: \"kubernetes.io/projected/a47abba7-dce0-433f-a78e-1b3224e07835-kube-api-access-kq7mc\") pod \"nova-scheduler-0\" (UID: \"a47abba7-dce0-433f-a78e-1b3224e07835\") " pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.702831 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:27:35 crc kubenswrapper[4735]: I1122 08:27:35.780776 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-72gww"] Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.211043 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.224433 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:27:36 crc kubenswrapper[4735]: W1122 08:27:36.234822 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda47abba7_dce0_433f_a78e_1b3224e07835.slice/crio-80ebe3a6a86eb2427b6c44748e56776aeb55d81b231bbde58bcfee54fa4c0705 WatchSource:0}: Error finding container 80ebe3a6a86eb2427b6c44748e56776aeb55d81b231bbde58bcfee54fa4c0705: Status 404 returned error can't find the container with id 80ebe3a6a86eb2427b6c44748e56776aeb55d81b231bbde58bcfee54fa4c0705 Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.323076 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhql7\" (UniqueName: \"kubernetes.io/projected/049fb0cb-f818-4234-a5ee-3bbfb518aec8-kube-api-access-rhql7\") pod \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.326953 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/049fb0cb-f818-4234-a5ee-3bbfb518aec8-kube-api-access-rhql7" (OuterVolumeSpecName: "kube-api-access-rhql7") pod "049fb0cb-f818-4234-a5ee-3bbfb518aec8" (UID: "049fb0cb-f818-4234-a5ee-3bbfb518aec8"). InnerVolumeSpecName "kube-api-access-rhql7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.336775 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-combined-ca-bundle\") pod \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.336875 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-config-data\") pod \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.336956 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-nova-metadata-tls-certs\") pod \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.337052 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/049fb0cb-f818-4234-a5ee-3bbfb518aec8-logs\") pod \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\" (UID: \"049fb0cb-f818-4234-a5ee-3bbfb518aec8\") " Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.347959 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/049fb0cb-f818-4234-a5ee-3bbfb518aec8-logs" (OuterVolumeSpecName: "logs") pod "049fb0cb-f818-4234-a5ee-3bbfb518aec8" (UID: "049fb0cb-f818-4234-a5ee-3bbfb518aec8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.349035 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/049fb0cb-f818-4234-a5ee-3bbfb518aec8-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.351667 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhql7\" (UniqueName: \"kubernetes.io/projected/049fb0cb-f818-4234-a5ee-3bbfb518aec8-kube-api-access-rhql7\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.355316 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a47abba7-dce0-433f-a78e-1b3224e07835","Type":"ContainerStarted","Data":"80ebe3a6a86eb2427b6c44748e56776aeb55d81b231bbde58bcfee54fa4c0705"} Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.417834 4735 generic.go:334] "Generic (PLEG): container finished" podID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerID="476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6" exitCode=0 Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.417922 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerDied","Data":"476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6"} Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.443393 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-72gww" event={"ID":"0365659b-65fd-4e91-a5bc-e49007a221b2","Type":"ContainerStarted","Data":"adcd28554860d4180448e2f1035cdbb99c2d29ceca838f4e86d58d0b01376da1"} Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.449126 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerStarted","Data":"2f363870df73f338b0337be8bbae9ce32d3c2bcfca834be0ab44e8d9121bea64"} Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.456071 4735 generic.go:334] "Generic (PLEG): container finished" podID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerID="65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798" exitCode=0 Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.456117 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"049fb0cb-f818-4234-a5ee-3bbfb518aec8","Type":"ContainerDied","Data":"65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798"} Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.456146 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"049fb0cb-f818-4234-a5ee-3bbfb518aec8","Type":"ContainerDied","Data":"89b3b94cd5481689d2dddc2cc71c04cfe5f2522a83b5b0fe0ed91963d4c3c50d"} Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.456166 4735 scope.go:117] "RemoveContainer" containerID="65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.456585 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.488601 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "049fb0cb-f818-4234-a5ee-3bbfb518aec8" (UID: "049fb0cb-f818-4234-a5ee-3bbfb518aec8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.497663 4735 scope.go:117] "RemoveContainer" containerID="7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.530712 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-config-data" (OuterVolumeSpecName: "config-data") pod "049fb0cb-f818-4234-a5ee-3bbfb518aec8" (UID: "049fb0cb-f818-4234-a5ee-3bbfb518aec8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.547126 4735 scope.go:117] "RemoveContainer" containerID="65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798" Nov 22 08:27:36 crc kubenswrapper[4735]: E1122 08:27:36.547687 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798\": container with ID starting with 65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798 not found: ID does not exist" containerID="65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.547729 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798"} err="failed to get container status \"65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798\": rpc error: code = NotFound desc = could not find container \"65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798\": container with ID starting with 65533ecb30e5d6d47b5a4681d60c68f8b8c269d82ebf7564cafed650f5547798 not found: ID does not exist" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.547756 4735 scope.go:117] "RemoveContainer" containerID="7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7" Nov 22 08:27:36 crc kubenswrapper[4735]: E1122 08:27:36.548173 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7\": container with ID starting with 7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7 not found: ID does not exist" containerID="7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.548223 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7"} err="failed to get container status \"7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7\": rpc error: code = NotFound desc = could not find container \"7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7\": container with ID starting with 7ef5ba6cf02b5275490cef0fa04d16af7f7081492c11a265cf9b0792cd3ec6e7 not found: ID does not exist" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.559428 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.559800 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.565362 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "049fb0cb-f818-4234-a5ee-3bbfb518aec8" (UID: "049fb0cb-f818-4234-a5ee-3bbfb518aec8"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.662520 4735 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/049fb0cb-f818-4234-a5ee-3bbfb518aec8-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.884498 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.893920 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.905532 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:27:36 crc kubenswrapper[4735]: E1122 08:27:36.906368 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-metadata" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.906452 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-metadata" Nov 22 08:27:36 crc kubenswrapper[4735]: E1122 08:27:36.906549 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-log" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.906615 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-log" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.906970 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-metadata" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.907048 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" containerName="nova-metadata-log" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.908873 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.914663 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.915851 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.916491 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.971719 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.972146 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.972260 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c346a589-fa62-4d07-9fcb-c4afb7f52b45-logs\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.972357 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-config-data\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:36 crc kubenswrapper[4735]: I1122 08:27:36.972603 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj6ks\" (UniqueName: \"kubernetes.io/projected/c346a589-fa62-4d07-9fcb-c4afb7f52b45-kube-api-access-xj6ks\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.074918 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.074984 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c346a589-fa62-4d07-9fcb-c4afb7f52b45-logs\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.075012 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-config-data\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.075041 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj6ks\" (UniqueName: \"kubernetes.io/projected/c346a589-fa62-4d07-9fcb-c4afb7f52b45-kube-api-access-xj6ks\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.075286 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.076245 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c346a589-fa62-4d07-9fcb-c4afb7f52b45-logs\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.079649 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.087881 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.088611 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c346a589-fa62-4d07-9fcb-c4afb7f52b45-config-data\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.109027 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj6ks\" (UniqueName: \"kubernetes.io/projected/c346a589-fa62-4d07-9fcb-c4afb7f52b45-kube-api-access-xj6ks\") pod \"nova-metadata-0\" (UID: \"c346a589-fa62-4d07-9fcb-c4afb7f52b45\") " pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.237230 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.282665 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="049fb0cb-f818-4234-a5ee-3bbfb518aec8" path="/var/lib/kubelet/pods/049fb0cb-f818-4234-a5ee-3bbfb518aec8/volumes" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.283802 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35" path="/var/lib/kubelet/pods/b1a4cf71-e395-4e1c-9ba7-1a3b37c99b35/volumes" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.471633 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerStarted","Data":"156ac2f42f920a1a333bc28b2ae024f5f7088f40be66c08f3ffbcc665b865d60"} Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.472028 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerStarted","Data":"a1d7ffb7b045a61e047393f56dc98e283830cfb2ad0de2b568dd67a6d6fcd433"} Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.473661 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a47abba7-dce0-433f-a78e-1b3224e07835","Type":"ContainerStarted","Data":"9a772d74d392980369c6ef43e35718970c853aed75ba83895719114ca88c1dac"} Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.479097 4735 generic.go:334] "Generic (PLEG): container finished" podID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerID="23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb" exitCode=0 Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.479146 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-72gww" event={"ID":"0365659b-65fd-4e91-a5bc-e49007a221b2","Type":"ContainerDied","Data":"23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb"} Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.491780 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.49176198 podStartE2EDuration="2.49176198s" podCreationTimestamp="2025-11-22 08:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:27:37.489661613 +0000 UTC m=+1479.094000218" watchObservedRunningTime="2025-11-22 08:27:37.49176198 +0000 UTC m=+1479.096100585" Nov 22 08:27:37 crc kubenswrapper[4735]: I1122 08:27:37.733630 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:27:37 crc kubenswrapper[4735]: W1122 08:27:37.735959 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc346a589_fa62_4d07_9fcb_c4afb7f52b45.slice/crio-ac61980c5fc49304e94db26c9bf104d3298fd16822fdaecb24070d0f82d3eaed WatchSource:0}: Error finding container ac61980c5fc49304e94db26c9bf104d3298fd16822fdaecb24070d0f82d3eaed: Status 404 returned error can't find the container with id ac61980c5fc49304e94db26c9bf104d3298fd16822fdaecb24070d0f82d3eaed Nov 22 08:27:38 crc kubenswrapper[4735]: I1122 08:27:38.492618 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c346a589-fa62-4d07-9fcb-c4afb7f52b45","Type":"ContainerStarted","Data":"d2d45a50df0c835b8142b377050f3f81c0ec9128142b6e128fec970063c4f150"} Nov 22 08:27:38 crc kubenswrapper[4735]: I1122 08:27:38.492665 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c346a589-fa62-4d07-9fcb-c4afb7f52b45","Type":"ContainerStarted","Data":"66974fbd094564626b7934a819fd089493bd319c10d6a3125f4bb99ead1a7823"} Nov 22 08:27:38 crc kubenswrapper[4735]: I1122 08:27:38.492681 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c346a589-fa62-4d07-9fcb-c4afb7f52b45","Type":"ContainerStarted","Data":"ac61980c5fc49304e94db26c9bf104d3298fd16822fdaecb24070d0f82d3eaed"} Nov 22 08:27:38 crc kubenswrapper[4735]: I1122 08:27:38.494325 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-72gww" event={"ID":"0365659b-65fd-4e91-a5bc-e49007a221b2","Type":"ContainerStarted","Data":"62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d"} Nov 22 08:27:38 crc kubenswrapper[4735]: I1122 08:27:38.498741 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerStarted","Data":"eb2196fe673298fd8e33ae46b0df1a1c75009e33f8c7ed31adff4ceefea57f4d"} Nov 22 08:27:38 crc kubenswrapper[4735]: I1122 08:27:38.521579 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.521560468 podStartE2EDuration="2.521560468s" podCreationTimestamp="2025-11-22 08:27:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:27:38.519731867 +0000 UTC m=+1480.124070482" watchObservedRunningTime="2025-11-22 08:27:38.521560468 +0000 UTC m=+1480.125899083" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.399238 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.516330 4735 generic.go:334] "Generic (PLEG): container finished" podID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerID="62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d" exitCode=0 Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.516410 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-72gww" event={"ID":"0365659b-65fd-4e91-a5bc-e49007a221b2","Type":"ContainerDied","Data":"62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d"} Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.520676 4735 generic.go:334] "Generic (PLEG): container finished" podID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerID="698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a" exitCode=0 Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.520737 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.520738 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc29c3b-bffc-4aac-81ff-d26313da9c78","Type":"ContainerDied","Data":"698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a"} Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.520769 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc29c3b-bffc-4aac-81ff-d26313da9c78","Type":"ContainerDied","Data":"9bc6da45354599733789f71ea347b9d17e5bc7d14a22fac81b9dcd4ae1c733bf"} Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.520784 4735 scope.go:117] "RemoveContainer" containerID="698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.543391 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgrg2\" (UniqueName: \"kubernetes.io/projected/efc29c3b-bffc-4aac-81ff-d26313da9c78-kube-api-access-cgrg2\") pod \"efc29c3b-bffc-4aac-81ff-d26313da9c78\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.543435 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-internal-tls-certs\") pod \"efc29c3b-bffc-4aac-81ff-d26313da9c78\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.543492 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-public-tls-certs\") pod \"efc29c3b-bffc-4aac-81ff-d26313da9c78\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.543578 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc29c3b-bffc-4aac-81ff-d26313da9c78-logs\") pod \"efc29c3b-bffc-4aac-81ff-d26313da9c78\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.543738 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-combined-ca-bundle\") pod \"efc29c3b-bffc-4aac-81ff-d26313da9c78\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.543786 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-config-data\") pod \"efc29c3b-bffc-4aac-81ff-d26313da9c78\" (UID: \"efc29c3b-bffc-4aac-81ff-d26313da9c78\") " Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.545570 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efc29c3b-bffc-4aac-81ff-d26313da9c78-logs" (OuterVolumeSpecName: "logs") pod "efc29c3b-bffc-4aac-81ff-d26313da9c78" (UID: "efc29c3b-bffc-4aac-81ff-d26313da9c78"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.556675 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efc29c3b-bffc-4aac-81ff-d26313da9c78-kube-api-access-cgrg2" (OuterVolumeSpecName: "kube-api-access-cgrg2") pod "efc29c3b-bffc-4aac-81ff-d26313da9c78" (UID: "efc29c3b-bffc-4aac-81ff-d26313da9c78"). InnerVolumeSpecName "kube-api-access-cgrg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.603025 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efc29c3b-bffc-4aac-81ff-d26313da9c78" (UID: "efc29c3b-bffc-4aac-81ff-d26313da9c78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.603440 4735 scope.go:117] "RemoveContainer" containerID="7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.614630 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-config-data" (OuterVolumeSpecName: "config-data") pod "efc29c3b-bffc-4aac-81ff-d26313da9c78" (UID: "efc29c3b-bffc-4aac-81ff-d26313da9c78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.625945 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "efc29c3b-bffc-4aac-81ff-d26313da9c78" (UID: "efc29c3b-bffc-4aac-81ff-d26313da9c78"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.646319 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.646488 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.646556 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgrg2\" (UniqueName: \"kubernetes.io/projected/efc29c3b-bffc-4aac-81ff-d26313da9c78-kube-api-access-cgrg2\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.646613 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.646665 4735 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc29c3b-bffc-4aac-81ff-d26313da9c78-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.652001 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "efc29c3b-bffc-4aac-81ff-d26313da9c78" (UID: "efc29c3b-bffc-4aac-81ff-d26313da9c78"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.678444 4735 scope.go:117] "RemoveContainer" containerID="698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a" Nov 22 08:27:39 crc kubenswrapper[4735]: E1122 08:27:39.678954 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a\": container with ID starting with 698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a not found: ID does not exist" containerID="698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.678988 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a"} err="failed to get container status \"698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a\": rpc error: code = NotFound desc = could not find container \"698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a\": container with ID starting with 698d58e03bc0ac01d38a519517cf0ba36d9533b290888291210c6e4ade5f5e0a not found: ID does not exist" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.679013 4735 scope.go:117] "RemoveContainer" containerID="7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9" Nov 22 08:27:39 crc kubenswrapper[4735]: E1122 08:27:39.679393 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9\": container with ID starting with 7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9 not found: ID does not exist" containerID="7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.679419 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9"} err="failed to get container status \"7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9\": rpc error: code = NotFound desc = could not find container \"7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9\": container with ID starting with 7454ec54255e1ffecde7750c5e42ac3f1ee45df8536b14189736000b93db9dd9 not found: ID does not exist" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.749335 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/efc29c3b-bffc-4aac-81ff-d26313da9c78-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.861351 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.873642 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.884077 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:39 crc kubenswrapper[4735]: E1122 08:27:39.888789 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-log" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.888830 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-log" Nov 22 08:27:39 crc kubenswrapper[4735]: E1122 08:27:39.888864 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-api" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.888873 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-api" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.889263 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-log" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.889295 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" containerName="nova-api-api" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.890846 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.898800 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.900064 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.900090 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.903896 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.956612 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-config-data\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.956659 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvh99\" (UniqueName: \"kubernetes.io/projected/14026d8e-f176-4c9f-8092-104e9a59ed5c-kube-api-access-bvh99\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.956686 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14026d8e-f176-4c9f-8092-104e9a59ed5c-logs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.956704 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.956738 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:39 crc kubenswrapper[4735]: I1122 08:27:39.956833 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.058794 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.059020 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-config-data\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.059055 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvh99\" (UniqueName: \"kubernetes.io/projected/14026d8e-f176-4c9f-8092-104e9a59ed5c-kube-api-access-bvh99\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.059104 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14026d8e-f176-4c9f-8092-104e9a59ed5c-logs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.059136 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.059193 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.059637 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14026d8e-f176-4c9f-8092-104e9a59ed5c-logs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.063950 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-config-data\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.065016 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.065306 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.078189 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14026d8e-f176-4c9f-8092-104e9a59ed5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.085817 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvh99\" (UniqueName: \"kubernetes.io/projected/14026d8e-f176-4c9f-8092-104e9a59ed5c-kube-api-access-bvh99\") pod \"nova-api-0\" (UID: \"14026d8e-f176-4c9f-8092-104e9a59ed5c\") " pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.218190 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.539672 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-72gww" event={"ID":"0365659b-65fd-4e91-a5bc-e49007a221b2","Type":"ContainerStarted","Data":"c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38"} Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.547747 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerStarted","Data":"28c47e499d47eba2a32f24c6d1eb33a2851c69c9376562d86e1b71e662ce0bb0"} Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.547938 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.561766 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-72gww" podStartSLOduration=4.019512685 podStartE2EDuration="6.561749661s" podCreationTimestamp="2025-11-22 08:27:34 +0000 UTC" firstStartedPulling="2025-11-22 08:27:37.481681264 +0000 UTC m=+1479.086019869" lastFinishedPulling="2025-11-22 08:27:40.02391824 +0000 UTC m=+1481.628256845" observedRunningTime="2025-11-22 08:27:40.555003316 +0000 UTC m=+1482.159341911" watchObservedRunningTime="2025-11-22 08:27:40.561749661 +0000 UTC m=+1482.166088266" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.590094 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.467403793 podStartE2EDuration="6.590067185s" podCreationTimestamp="2025-11-22 08:27:34 +0000 UTC" firstStartedPulling="2025-11-22 08:27:35.482890034 +0000 UTC m=+1477.087228639" lastFinishedPulling="2025-11-22 08:27:39.605553426 +0000 UTC m=+1481.209892031" observedRunningTime="2025-11-22 08:27:40.577667436 +0000 UTC m=+1482.182006041" watchObservedRunningTime="2025-11-22 08:27:40.590067185 +0000 UTC m=+1482.194405790" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.703920 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 08:27:40 crc kubenswrapper[4735]: I1122 08:27:40.758511 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:27:40 crc kubenswrapper[4735]: W1122 08:27:40.768041 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14026d8e_f176_4c9f_8092_104e9a59ed5c.slice/crio-ef13fda51506eeb4ca20f0051b856c9ed5134066f67d78acaf56e36eb0c10bf9 WatchSource:0}: Error finding container ef13fda51506eeb4ca20f0051b856c9ed5134066f67d78acaf56e36eb0c10bf9: Status 404 returned error can't find the container with id ef13fda51506eeb4ca20f0051b856c9ed5134066f67d78acaf56e36eb0c10bf9 Nov 22 08:27:41 crc kubenswrapper[4735]: I1122 08:27:41.275655 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efc29c3b-bffc-4aac-81ff-d26313da9c78" path="/var/lib/kubelet/pods/efc29c3b-bffc-4aac-81ff-d26313da9c78/volumes" Nov 22 08:27:41 crc kubenswrapper[4735]: I1122 08:27:41.614481 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14026d8e-f176-4c9f-8092-104e9a59ed5c","Type":"ContainerStarted","Data":"126e9ba76e77130234147a3a8a6d07b5fcf1e37afdc3bd46e4be8d5c8d4a4377"} Nov 22 08:27:41 crc kubenswrapper[4735]: I1122 08:27:41.614529 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14026d8e-f176-4c9f-8092-104e9a59ed5c","Type":"ContainerStarted","Data":"c754900b21dbb9184139d9260236a1d2d819fc74f7f34f72706a2c7aa9f5e47f"} Nov 22 08:27:41 crc kubenswrapper[4735]: I1122 08:27:41.614538 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14026d8e-f176-4c9f-8092-104e9a59ed5c","Type":"ContainerStarted","Data":"ef13fda51506eeb4ca20f0051b856c9ed5134066f67d78acaf56e36eb0c10bf9"} Nov 22 08:27:41 crc kubenswrapper[4735]: I1122 08:27:41.638946 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.638926643 podStartE2EDuration="2.638926643s" podCreationTimestamp="2025-11-22 08:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:27:41.633236748 +0000 UTC m=+1483.237575353" watchObservedRunningTime="2025-11-22 08:27:41.638926643 +0000 UTC m=+1483.243265248" Nov 22 08:27:42 crc kubenswrapper[4735]: I1122 08:27:42.239256 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:27:42 crc kubenswrapper[4735]: I1122 08:27:42.239313 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:27:45 crc kubenswrapper[4735]: I1122 08:27:45.095810 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:45 crc kubenswrapper[4735]: I1122 08:27:45.096286 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:45 crc kubenswrapper[4735]: I1122 08:27:45.151207 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:45 crc kubenswrapper[4735]: I1122 08:27:45.703943 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 22 08:27:45 crc kubenswrapper[4735]: I1122 08:27:45.761038 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:45 crc kubenswrapper[4735]: I1122 08:27:45.787728 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 22 08:27:45 crc kubenswrapper[4735]: I1122 08:27:45.837350 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-72gww"] Nov 22 08:27:46 crc kubenswrapper[4735]: I1122 08:27:46.131488 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:27:46 crc kubenswrapper[4735]: I1122 08:27:46.131603 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:27:46 crc kubenswrapper[4735]: I1122 08:27:46.732996 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 22 08:27:47 crc kubenswrapper[4735]: I1122 08:27:47.239057 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 08:27:47 crc kubenswrapper[4735]: I1122 08:27:47.239129 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 08:27:47 crc kubenswrapper[4735]: I1122 08:27:47.684275 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-72gww" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerName="registry-server" containerID="cri-o://c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38" gracePeriod=2 Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.251837 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.254654 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c346a589-fa62-4d07-9fcb-c4afb7f52b45" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.3:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.254658 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c346a589-fa62-4d07-9fcb-c4afb7f52b45" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.3:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.381840 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-catalog-content\") pod \"0365659b-65fd-4e91-a5bc-e49007a221b2\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.381909 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxlgp\" (UniqueName: \"kubernetes.io/projected/0365659b-65fd-4e91-a5bc-e49007a221b2-kube-api-access-bxlgp\") pod \"0365659b-65fd-4e91-a5bc-e49007a221b2\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.382111 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-utilities\") pod \"0365659b-65fd-4e91-a5bc-e49007a221b2\" (UID: \"0365659b-65fd-4e91-a5bc-e49007a221b2\") " Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.383125 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-utilities" (OuterVolumeSpecName: "utilities") pod "0365659b-65fd-4e91-a5bc-e49007a221b2" (UID: "0365659b-65fd-4e91-a5bc-e49007a221b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.388689 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0365659b-65fd-4e91-a5bc-e49007a221b2-kube-api-access-bxlgp" (OuterVolumeSpecName: "kube-api-access-bxlgp") pod "0365659b-65fd-4e91-a5bc-e49007a221b2" (UID: "0365659b-65fd-4e91-a5bc-e49007a221b2"). InnerVolumeSpecName "kube-api-access-bxlgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.405626 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0365659b-65fd-4e91-a5bc-e49007a221b2" (UID: "0365659b-65fd-4e91-a5bc-e49007a221b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.484777 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.484813 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0365659b-65fd-4e91-a5bc-e49007a221b2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.484824 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxlgp\" (UniqueName: \"kubernetes.io/projected/0365659b-65fd-4e91-a5bc-e49007a221b2-kube-api-access-bxlgp\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.697541 4735 generic.go:334] "Generic (PLEG): container finished" podID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerID="c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38" exitCode=0 Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.697587 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-72gww" event={"ID":"0365659b-65fd-4e91-a5bc-e49007a221b2","Type":"ContainerDied","Data":"c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38"} Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.697619 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-72gww" event={"ID":"0365659b-65fd-4e91-a5bc-e49007a221b2","Type":"ContainerDied","Data":"adcd28554860d4180448e2f1035cdbb99c2d29ceca838f4e86d58d0b01376da1"} Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.697622 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-72gww" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.697638 4735 scope.go:117] "RemoveContainer" containerID="c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.723630 4735 scope.go:117] "RemoveContainer" containerID="62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.740313 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-72gww"] Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.772481 4735 scope.go:117] "RemoveContainer" containerID="23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.805423 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-72gww"] Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.816786 4735 scope.go:117] "RemoveContainer" containerID="c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38" Nov 22 08:27:48 crc kubenswrapper[4735]: E1122 08:27:48.817163 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38\": container with ID starting with c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38 not found: ID does not exist" containerID="c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.817193 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38"} err="failed to get container status \"c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38\": rpc error: code = NotFound desc = could not find container \"c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38\": container with ID starting with c1c516d600f2766e0d434dfba053c1818be1ba095a8f738f536433fa1988fc38 not found: ID does not exist" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.817215 4735 scope.go:117] "RemoveContainer" containerID="62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d" Nov 22 08:27:48 crc kubenswrapper[4735]: E1122 08:27:48.817706 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d\": container with ID starting with 62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d not found: ID does not exist" containerID="62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.819489 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d"} err="failed to get container status \"62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d\": rpc error: code = NotFound desc = could not find container \"62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d\": container with ID starting with 62d85fafe290b810631e68fc8c72d07a19106c8376cde47058e4a1222fc2734d not found: ID does not exist" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.819536 4735 scope.go:117] "RemoveContainer" containerID="23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb" Nov 22 08:27:48 crc kubenswrapper[4735]: E1122 08:27:48.819821 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb\": container with ID starting with 23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb not found: ID does not exist" containerID="23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb" Nov 22 08:27:48 crc kubenswrapper[4735]: I1122 08:27:48.819846 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb"} err="failed to get container status \"23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb\": rpc error: code = NotFound desc = could not find container \"23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb\": container with ID starting with 23bafc5b84f65469e66825ba04c500c117f3be21db49a5a3a81f3f7e7dd988fb not found: ID does not exist" Nov 22 08:27:49 crc kubenswrapper[4735]: I1122 08:27:49.281765 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" path="/var/lib/kubelet/pods/0365659b-65fd-4e91-a5bc-e49007a221b2/volumes" Nov 22 08:27:49 crc kubenswrapper[4735]: I1122 08:27:49.693133 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="69fdd40e-0c07-49c4-a65f-e5be20afbf5d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.245:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:27:50 crc kubenswrapper[4735]: I1122 08:27:50.218398 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:27:50 crc kubenswrapper[4735]: I1122 08:27:50.218444 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:27:51 crc kubenswrapper[4735]: I1122 08:27:51.239582 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="14026d8e-f176-4c9f-8092-104e9a59ed5c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.4:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:27:51 crc kubenswrapper[4735]: I1122 08:27:51.239604 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="14026d8e-f176-4c9f-8092-104e9a59ed5c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.4:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:27:57 crc kubenswrapper[4735]: I1122 08:27:57.249798 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 08:27:57 crc kubenswrapper[4735]: I1122 08:27:57.250628 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 08:27:57 crc kubenswrapper[4735]: I1122 08:27:57.259940 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 08:27:57 crc kubenswrapper[4735]: I1122 08:27:57.281785 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 08:28:00 crc kubenswrapper[4735]: I1122 08:28:00.224417 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 08:28:00 crc kubenswrapper[4735]: I1122 08:28:00.225306 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 08:28:00 crc kubenswrapper[4735]: I1122 08:28:00.226236 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 08:28:00 crc kubenswrapper[4735]: I1122 08:28:00.226271 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 08:28:00 crc kubenswrapper[4735]: I1122 08:28:00.232672 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 08:28:00 crc kubenswrapper[4735]: I1122 08:28:00.232985 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.681213 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.682609 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.782843 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjq6t\" (UniqueName: \"kubernetes.io/projected/477b15d5-59f1-4f08-97eb-14ae37feb4ae-kube-api-access-fjq6t\") pod \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.783175 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-config-data\") pod \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.783255 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-scripts\") pod \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.783406 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-combined-ca-bundle\") pod \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\" (UID: \"477b15d5-59f1-4f08-97eb-14ae37feb4ae\") " Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.789112 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/477b15d5-59f1-4f08-97eb-14ae37feb4ae-kube-api-access-fjq6t" (OuterVolumeSpecName: "kube-api-access-fjq6t") pod "477b15d5-59f1-4f08-97eb-14ae37feb4ae" (UID: "477b15d5-59f1-4f08-97eb-14ae37feb4ae"). InnerVolumeSpecName "kube-api-access-fjq6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.801683 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-scripts" (OuterVolumeSpecName: "scripts") pod "477b15d5-59f1-4f08-97eb-14ae37feb4ae" (UID: "477b15d5-59f1-4f08-97eb-14ae37feb4ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.888965 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.889185 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjq6t\" (UniqueName: \"kubernetes.io/projected/477b15d5-59f1-4f08-97eb-14ae37feb4ae-kube-api-access-fjq6t\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.912533 4735 generic.go:334] "Generic (PLEG): container finished" podID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerID="af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a" exitCode=137 Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.912578 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerDied","Data":"af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a"} Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.912605 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"477b15d5-59f1-4f08-97eb-14ae37feb4ae","Type":"ContainerDied","Data":"62182e2ac75e8783c7308cb955c32951edfbd0ee45431198ef15199de82f2915"} Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.912621 4735 scope.go:117] "RemoveContainer" containerID="af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.912788 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.928431 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-config-data" (OuterVolumeSpecName: "config-data") pod "477b15d5-59f1-4f08-97eb-14ae37feb4ae" (UID: "477b15d5-59f1-4f08-97eb-14ae37feb4ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.937171 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "477b15d5-59f1-4f08-97eb-14ae37feb4ae" (UID: "477b15d5-59f1-4f08-97eb-14ae37feb4ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.958269 4735 scope.go:117] "RemoveContainer" containerID="476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.983650 4735 scope.go:117] "RemoveContainer" containerID="90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.991475 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:04 crc kubenswrapper[4735]: I1122 08:28:04.991504 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/477b15d5-59f1-4f08-97eb-14ae37feb4ae-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.008589 4735 scope.go:117] "RemoveContainer" containerID="13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.044203 4735 scope.go:117] "RemoveContainer" containerID="af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.044632 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a\": container with ID starting with af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a not found: ID does not exist" containerID="af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.044660 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a"} err="failed to get container status \"af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a\": rpc error: code = NotFound desc = could not find container \"af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a\": container with ID starting with af008cf7897b029e485278b9b492e0154ce39ccf7c639c5efa390aca746cfa2a not found: ID does not exist" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.044680 4735 scope.go:117] "RemoveContainer" containerID="476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.044944 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6\": container with ID starting with 476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6 not found: ID does not exist" containerID="476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.044970 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6"} err="failed to get container status \"476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6\": rpc error: code = NotFound desc = could not find container \"476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6\": container with ID starting with 476a402fb5e7f60a3731f66e4124cbf51cc7b77e2b804c433dab897a2123d3d6 not found: ID does not exist" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.044985 4735 scope.go:117] "RemoveContainer" containerID="90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.045243 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e\": container with ID starting with 90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e not found: ID does not exist" containerID="90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.045264 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e"} err="failed to get container status \"90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e\": rpc error: code = NotFound desc = could not find container \"90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e\": container with ID starting with 90d2153deef951b932fda349c38c2a85a689a42b52951651142d67fe6df4033e not found: ID does not exist" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.045277 4735 scope.go:117] "RemoveContainer" containerID="13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.045567 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d\": container with ID starting with 13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d not found: ID does not exist" containerID="13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.045590 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d"} err="failed to get container status \"13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d\": rpc error: code = NotFound desc = could not find container \"13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d\": container with ID starting with 13c9c3dcd7b3f1ef4dbf3ecaa58ed9a82c3c76fec969792678ed423d18f4b60d not found: ID does not exist" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.285794 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.285946 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.305441 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.305930 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-listener" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.305947 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-listener" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.305964 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerName="extract-utilities" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.305973 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerName="extract-utilities" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.305995 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-notifier" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306002 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-notifier" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.306024 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-evaluator" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306030 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-evaluator" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.306042 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-api" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306067 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-api" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.306093 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerName="extract-content" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306100 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerName="extract-content" Nov 22 08:28:05 crc kubenswrapper[4735]: E1122 08:28:05.306111 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerName="registry-server" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306117 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerName="registry-server" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306316 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-api" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306339 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-evaluator" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306356 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-notifier" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306373 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" containerName="aodh-listener" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.306383 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0365659b-65fd-4e91-a5bc-e49007a221b2" containerName="registry-server" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.308587 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.310908 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.316758 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.317732 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.318703 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-bfgs7" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.319747 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.322521 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.407410 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-scripts\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.407803 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-public-tls-certs\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.407901 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mdb7\" (UniqueName: \"kubernetes.io/projected/72122e00-8d36-467a-9fca-d33d73ef458b-kube-api-access-8mdb7\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.407976 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-internal-tls-certs\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.407996 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-config-data\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.408024 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.510123 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-scripts\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.510271 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-public-tls-certs\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.510375 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mdb7\" (UniqueName: \"kubernetes.io/projected/72122e00-8d36-467a-9fca-d33d73ef458b-kube-api-access-8mdb7\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.510436 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-internal-tls-certs\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.510909 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-config-data\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.510939 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.513704 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-internal-tls-certs\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.514237 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-scripts\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.514355 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-config-data\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.514706 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.515931 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-public-tls-certs\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.527083 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mdb7\" (UniqueName: \"kubernetes.io/projected/72122e00-8d36-467a-9fca-d33d73ef458b-kube-api-access-8mdb7\") pod \"aodh-0\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " pod="openstack/aodh-0" Nov 22 08:28:05 crc kubenswrapper[4735]: I1122 08:28:05.670179 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:28:06 crc kubenswrapper[4735]: I1122 08:28:06.168018 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 22 08:28:06 crc kubenswrapper[4735]: W1122 08:28:06.179620 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72122e00_8d36_467a_9fca_d33d73ef458b.slice/crio-108b387a0f538fdc450cfa4822a53dfabc4a43f4e31519c5079782b4b062fc70 WatchSource:0}: Error finding container 108b387a0f538fdc450cfa4822a53dfabc4a43f4e31519c5079782b4b062fc70: Status 404 returned error can't find the container with id 108b387a0f538fdc450cfa4822a53dfabc4a43f4e31519c5079782b4b062fc70 Nov 22 08:28:06 crc kubenswrapper[4735]: I1122 08:28:06.939561 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerStarted","Data":"108b387a0f538fdc450cfa4822a53dfabc4a43f4e31519c5079782b4b062fc70"} Nov 22 08:28:07 crc kubenswrapper[4735]: I1122 08:28:07.279036 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="477b15d5-59f1-4f08-97eb-14ae37feb4ae" path="/var/lib/kubelet/pods/477b15d5-59f1-4f08-97eb-14ae37feb4ae/volumes" Nov 22 08:28:07 crc kubenswrapper[4735]: I1122 08:28:07.956452 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerStarted","Data":"0798dc5eda46a091c75461a2bb9584d339a13a861f521d088d8295c91440a1d4"} Nov 22 08:28:08 crc kubenswrapper[4735]: I1122 08:28:08.978247 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerStarted","Data":"ae00901a5c9b52914ee48e511f5380bb1657eb56dee8e3af32ea8fce46538b32"} Nov 22 08:28:08 crc kubenswrapper[4735]: I1122 08:28:08.979931 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerStarted","Data":"e8cbdd89932353d5e2975889e319e29e0af3f59f699ee85b488473c10d8fba74"} Nov 22 08:28:09 crc kubenswrapper[4735]: I1122 08:28:09.249319 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:28:09 crc kubenswrapper[4735]: I1122 08:28:09.249589 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="4c7911f8-389a-43fe-a011-dcc65ccbd7f2" containerName="kube-state-metrics" containerID="cri-o://786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9" gracePeriod=30 Nov 22 08:28:09 crc kubenswrapper[4735]: I1122 08:28:09.378072 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:28:09 crc kubenswrapper[4735]: I1122 08:28:09.378587 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="3b29201e-860c-4944-ad29-06abb7aed0c6" containerName="mysqld-exporter" containerID="cri-o://12ecbfb69646afa24be66e782ce2df4e41d0007a79333f6206eaf9333ca628a4" gracePeriod=30 Nov 22 08:28:09 crc kubenswrapper[4735]: I1122 08:28:09.878918 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.050657 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgmkz\" (UniqueName: \"kubernetes.io/projected/4c7911f8-389a-43fe-a011-dcc65ccbd7f2-kube-api-access-dgmkz\") pod \"4c7911f8-389a-43fe-a011-dcc65ccbd7f2\" (UID: \"4c7911f8-389a-43fe-a011-dcc65ccbd7f2\") " Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.074183 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c7911f8-389a-43fe-a011-dcc65ccbd7f2-kube-api-access-dgmkz" (OuterVolumeSpecName: "kube-api-access-dgmkz") pod "4c7911f8-389a-43fe-a011-dcc65ccbd7f2" (UID: "4c7911f8-389a-43fe-a011-dcc65ccbd7f2"). InnerVolumeSpecName "kube-api-access-dgmkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.080551 4735 generic.go:334] "Generic (PLEG): container finished" podID="3b29201e-860c-4944-ad29-06abb7aed0c6" containerID="12ecbfb69646afa24be66e782ce2df4e41d0007a79333f6206eaf9333ca628a4" exitCode=2 Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.080633 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"3b29201e-860c-4944-ad29-06abb7aed0c6","Type":"ContainerDied","Data":"12ecbfb69646afa24be66e782ce2df4e41d0007a79333f6206eaf9333ca628a4"} Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.082651 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerStarted","Data":"6462b6967208bd3cca5d8300b6b254224c25e883fc9eaa70d61b234d15c25876"} Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.118897 4735 generic.go:334] "Generic (PLEG): container finished" podID="4c7911f8-389a-43fe-a011-dcc65ccbd7f2" containerID="786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9" exitCode=2 Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.118968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4c7911f8-389a-43fe-a011-dcc65ccbd7f2","Type":"ContainerDied","Data":"786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9"} Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.118995 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4c7911f8-389a-43fe-a011-dcc65ccbd7f2","Type":"ContainerDied","Data":"21fdf04e69138814f021a78e2b82930f926643c1eacc99f95d0811e5978b4e63"} Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.119011 4735 scope.go:117] "RemoveContainer" containerID="786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.119218 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.138095 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.004964474 podStartE2EDuration="5.138078787s" podCreationTimestamp="2025-11-22 08:28:05 +0000 UTC" firstStartedPulling="2025-11-22 08:28:06.182384285 +0000 UTC m=+1507.786722890" lastFinishedPulling="2025-11-22 08:28:09.315498598 +0000 UTC m=+1510.919837203" observedRunningTime="2025-11-22 08:28:10.136940355 +0000 UTC m=+1511.741278960" watchObservedRunningTime="2025-11-22 08:28:10.138078787 +0000 UTC m=+1511.742417382" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.162134 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgmkz\" (UniqueName: \"kubernetes.io/projected/4c7911f8-389a-43fe-a011-dcc65ccbd7f2-kube-api-access-dgmkz\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.205494 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.208681 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.219885 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.231751 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:28:10 crc kubenswrapper[4735]: E1122 08:28:10.232205 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b29201e-860c-4944-ad29-06abb7aed0c6" containerName="mysqld-exporter" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.232222 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b29201e-860c-4944-ad29-06abb7aed0c6" containerName="mysqld-exporter" Nov 22 08:28:10 crc kubenswrapper[4735]: E1122 08:28:10.232243 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7911f8-389a-43fe-a011-dcc65ccbd7f2" containerName="kube-state-metrics" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.232268 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7911f8-389a-43fe-a011-dcc65ccbd7f2" containerName="kube-state-metrics" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.232521 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b29201e-860c-4944-ad29-06abb7aed0c6" containerName="mysqld-exporter" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.232539 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c7911f8-389a-43fe-a011-dcc65ccbd7f2" containerName="kube-state-metrics" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.233302 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.240443 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.240665 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.297752 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.328731 4735 scope.go:117] "RemoveContainer" containerID="786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9" Nov 22 08:28:10 crc kubenswrapper[4735]: E1122 08:28:10.332839 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9\": container with ID starting with 786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9 not found: ID does not exist" containerID="786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.332875 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9"} err="failed to get container status \"786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9\": rpc error: code = NotFound desc = could not find container \"786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9\": container with ID starting with 786bff82303083a2c9249409650105aa16f100d7eba1ae84e076a97c91f56ce9 not found: ID does not exist" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.368972 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f8sz\" (UniqueName: \"kubernetes.io/projected/3b29201e-860c-4944-ad29-06abb7aed0c6-kube-api-access-6f8sz\") pod \"3b29201e-860c-4944-ad29-06abb7aed0c6\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.369077 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-combined-ca-bundle\") pod \"3b29201e-860c-4944-ad29-06abb7aed0c6\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.369279 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-config-data\") pod \"3b29201e-860c-4944-ad29-06abb7aed0c6\" (UID: \"3b29201e-860c-4944-ad29-06abb7aed0c6\") " Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.371232 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.371376 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.375650 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.375819 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49mvc\" (UniqueName: \"kubernetes.io/projected/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-api-access-49mvc\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.379957 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b29201e-860c-4944-ad29-06abb7aed0c6-kube-api-access-6f8sz" (OuterVolumeSpecName: "kube-api-access-6f8sz") pod "3b29201e-860c-4944-ad29-06abb7aed0c6" (UID: "3b29201e-860c-4944-ad29-06abb7aed0c6"). InnerVolumeSpecName "kube-api-access-6f8sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.419775 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b29201e-860c-4944-ad29-06abb7aed0c6" (UID: "3b29201e-860c-4944-ad29-06abb7aed0c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.470077 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-config-data" (OuterVolumeSpecName: "config-data") pod "3b29201e-860c-4944-ad29-06abb7aed0c6" (UID: "3b29201e-860c-4944-ad29-06abb7aed0c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.478027 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.478149 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.478215 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49mvc\" (UniqueName: \"kubernetes.io/projected/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-api-access-49mvc\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.478261 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.478330 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.478346 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b29201e-860c-4944-ad29-06abb7aed0c6-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.478361 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f8sz\" (UniqueName: \"kubernetes.io/projected/3b29201e-860c-4944-ad29-06abb7aed0c6-kube-api-access-6f8sz\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.486566 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.488162 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.490964 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378d8516-81a6-4d5a-99cd-d5c4190370af-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.497978 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49mvc\" (UniqueName: \"kubernetes.io/projected/378d8516-81a6-4d5a-99cd-d5c4190370af-kube-api-access-49mvc\") pod \"kube-state-metrics-0\" (UID: \"378d8516-81a6-4d5a-99cd-d5c4190370af\") " pod="openstack/kube-state-metrics-0" Nov 22 08:28:10 crc kubenswrapper[4735]: I1122 08:28:10.558134 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.130938 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"3b29201e-860c-4944-ad29-06abb7aed0c6","Type":"ContainerDied","Data":"4da5c46aa8080594d1437e5b10e078828e7b5a8c8bedb8c0c250d3eeb35cbedd"} Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.130988 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.131169 4735 scope.go:117] "RemoveContainer" containerID="12ecbfb69646afa24be66e782ce2df4e41d0007a79333f6206eaf9333ca628a4" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.188521 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.209660 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.228118 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.241119 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.242590 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.245932 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.255968 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.257988 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.305373 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b29201e-860c-4944-ad29-06abb7aed0c6" path="/var/lib/kubelet/pods/3b29201e-860c-4944-ad29-06abb7aed0c6/volumes" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.307212 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c7911f8-389a-43fe-a011-dcc65ccbd7f2" path="/var/lib/kubelet/pods/4c7911f8-389a-43fe-a011-dcc65ccbd7f2/volumes" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.410522 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v77r9\" (UniqueName: \"kubernetes.io/projected/da023172-a0d0-4420-be51-05b348cf6c0c-kube-api-access-v77r9\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.410690 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.410772 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-config-data\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.410800 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.512918 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.513090 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-config-data\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.513143 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.513239 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v77r9\" (UniqueName: \"kubernetes.io/projected/da023172-a0d0-4420-be51-05b348cf6c0c-kube-api-access-v77r9\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.520390 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-config-data\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.520715 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.527037 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/da023172-a0d0-4420-be51-05b348cf6c0c-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.532316 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v77r9\" (UniqueName: \"kubernetes.io/projected/da023172-a0d0-4420-be51-05b348cf6c0c-kube-api-access-v77r9\") pod \"mysqld-exporter-0\" (UID: \"da023172-a0d0-4420-be51-05b348cf6c0c\") " pod="openstack/mysqld-exporter-0" Nov 22 08:28:11 crc kubenswrapper[4735]: I1122 08:28:11.572030 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.101834 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 22 08:28:12 crc kubenswrapper[4735]: W1122 08:28:12.108839 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda023172_a0d0_4420_be51_05b348cf6c0c.slice/crio-222b9dee8d7c7479be66a0afa3137e86684eedf23691c68eba263fb7fb799fed WatchSource:0}: Error finding container 222b9dee8d7c7479be66a0afa3137e86684eedf23691c68eba263fb7fb799fed: Status 404 returned error can't find the container with id 222b9dee8d7c7479be66a0afa3137e86684eedf23691c68eba263fb7fb799fed Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.150131 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"da023172-a0d0-4420-be51-05b348cf6c0c","Type":"ContainerStarted","Data":"222b9dee8d7c7479be66a0afa3137e86684eedf23691c68eba263fb7fb799fed"} Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.164273 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"378d8516-81a6-4d5a-99cd-d5c4190370af","Type":"ContainerStarted","Data":"a254644cf26c70bd33aee9f90c256e8736a1bfdf3f41eb90893cadeaead7c898"} Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.164313 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"378d8516-81a6-4d5a-99cd-d5c4190370af","Type":"ContainerStarted","Data":"f5243aac2c6d47cfd72b5fc19c7def85d1acdf844ab393d4e9549d22f456b9b7"} Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.164434 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.194189 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.758545272 podStartE2EDuration="2.194167295s" podCreationTimestamp="2025-11-22 08:28:10 +0000 UTC" firstStartedPulling="2025-11-22 08:28:11.210333538 +0000 UTC m=+1512.814672143" lastFinishedPulling="2025-11-22 08:28:11.645955561 +0000 UTC m=+1513.250294166" observedRunningTime="2025-11-22 08:28:12.186623277 +0000 UTC m=+1513.790961892" watchObservedRunningTime="2025-11-22 08:28:12.194167295 +0000 UTC m=+1513.798505900" Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.346272 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.346558 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="ceilometer-central-agent" containerID="cri-o://a1d7ffb7b045a61e047393f56dc98e283830cfb2ad0de2b568dd67a6d6fcd433" gracePeriod=30 Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.346917 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="proxy-httpd" containerID="cri-o://28c47e499d47eba2a32f24c6d1eb33a2851c69c9376562d86e1b71e662ce0bb0" gracePeriod=30 Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.346971 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="ceilometer-notification-agent" containerID="cri-o://156ac2f42f920a1a333bc28b2ae024f5f7088f40be66c08f3ffbcc665b865d60" gracePeriod=30 Nov 22 08:28:12 crc kubenswrapper[4735]: I1122 08:28:12.346992 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="sg-core" containerID="cri-o://eb2196fe673298fd8e33ae46b0df1a1c75009e33f8c7ed31adff4ceefea57f4d" gracePeriod=30 Nov 22 08:28:13 crc kubenswrapper[4735]: I1122 08:28:13.180125 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"da023172-a0d0-4420-be51-05b348cf6c0c","Type":"ContainerStarted","Data":"b1e8ea04e9ab5f941dbfc8c70fd923cd880fca9f816c880c5b04538311aabe31"} Nov 22 08:28:13 crc kubenswrapper[4735]: I1122 08:28:13.184372 4735 generic.go:334] "Generic (PLEG): container finished" podID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerID="28c47e499d47eba2a32f24c6d1eb33a2851c69c9376562d86e1b71e662ce0bb0" exitCode=0 Nov 22 08:28:13 crc kubenswrapper[4735]: I1122 08:28:13.184436 4735 generic.go:334] "Generic (PLEG): container finished" podID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerID="eb2196fe673298fd8e33ae46b0df1a1c75009e33f8c7ed31adff4ceefea57f4d" exitCode=2 Nov 22 08:28:13 crc kubenswrapper[4735]: I1122 08:28:13.184497 4735 generic.go:334] "Generic (PLEG): container finished" podID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerID="a1d7ffb7b045a61e047393f56dc98e283830cfb2ad0de2b568dd67a6d6fcd433" exitCode=0 Nov 22 08:28:13 crc kubenswrapper[4735]: I1122 08:28:13.184434 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerDied","Data":"28c47e499d47eba2a32f24c6d1eb33a2851c69c9376562d86e1b71e662ce0bb0"} Nov 22 08:28:13 crc kubenswrapper[4735]: I1122 08:28:13.184978 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerDied","Data":"eb2196fe673298fd8e33ae46b0df1a1c75009e33f8c7ed31adff4ceefea57f4d"} Nov 22 08:28:13 crc kubenswrapper[4735]: I1122 08:28:13.185025 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerDied","Data":"a1d7ffb7b045a61e047393f56dc98e283830cfb2ad0de2b568dd67a6d6fcd433"} Nov 22 08:28:13 crc kubenswrapper[4735]: I1122 08:28:13.197188 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=1.657673708 podStartE2EDuration="2.197160952s" podCreationTimestamp="2025-11-22 08:28:11 +0000 UTC" firstStartedPulling="2025-11-22 08:28:12.112583129 +0000 UTC m=+1513.716921734" lastFinishedPulling="2025-11-22 08:28:12.652070373 +0000 UTC m=+1514.256408978" observedRunningTime="2025-11-22 08:28:13.193888551 +0000 UTC m=+1514.798227156" watchObservedRunningTime="2025-11-22 08:28:13.197160952 +0000 UTC m=+1514.801499587" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.202683 4735 generic.go:334] "Generic (PLEG): container finished" podID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerID="156ac2f42f920a1a333bc28b2ae024f5f7088f40be66c08f3ffbcc665b865d60" exitCode=0 Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.203514 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerDied","Data":"156ac2f42f920a1a333bc28b2ae024f5f7088f40be66c08f3ffbcc665b865d60"} Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.399234 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.436259 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-scripts\") pod \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.436423 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-combined-ca-bundle\") pod \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.436673 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-sg-core-conf-yaml\") pod \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.436802 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-config-data\") pod \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.436896 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-run-httpd\") pod \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.436943 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llnn9\" (UniqueName: \"kubernetes.io/projected/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-kube-api-access-llnn9\") pod \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.436989 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-log-httpd\") pod \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\" (UID: \"89192c39-d1e5-4ca8-a1f2-b4c976ae7802\") " Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.437817 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "89192c39-d1e5-4ca8-a1f2-b4c976ae7802" (UID: "89192c39-d1e5-4ca8-a1f2-b4c976ae7802"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.438201 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "89192c39-d1e5-4ca8-a1f2-b4c976ae7802" (UID: "89192c39-d1e5-4ca8-a1f2-b4c976ae7802"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.439501 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.439566 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.459253 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-kube-api-access-llnn9" (OuterVolumeSpecName: "kube-api-access-llnn9") pod "89192c39-d1e5-4ca8-a1f2-b4c976ae7802" (UID: "89192c39-d1e5-4ca8-a1f2-b4c976ae7802"). InnerVolumeSpecName "kube-api-access-llnn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.466655 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-scripts" (OuterVolumeSpecName: "scripts") pod "89192c39-d1e5-4ca8-a1f2-b4c976ae7802" (UID: "89192c39-d1e5-4ca8-a1f2-b4c976ae7802"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.519239 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "89192c39-d1e5-4ca8-a1f2-b4c976ae7802" (UID: "89192c39-d1e5-4ca8-a1f2-b4c976ae7802"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.543677 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.543724 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llnn9\" (UniqueName: \"kubernetes.io/projected/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-kube-api-access-llnn9\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.543738 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.627040 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89192c39-d1e5-4ca8-a1f2-b4c976ae7802" (UID: "89192c39-d1e5-4ca8-a1f2-b4c976ae7802"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.646200 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.658306 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-config-data" (OuterVolumeSpecName: "config-data") pod "89192c39-d1e5-4ca8-a1f2-b4c976ae7802" (UID: "89192c39-d1e5-4ca8-a1f2-b4c976ae7802"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:14 crc kubenswrapper[4735]: I1122 08:28:14.749153 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89192c39-d1e5-4ca8-a1f2-b4c976ae7802-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.215270 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89192c39-d1e5-4ca8-a1f2-b4c976ae7802","Type":"ContainerDied","Data":"2f363870df73f338b0337be8bbae9ce32d3c2bcfca834be0ab44e8d9121bea64"} Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.215560 4735 scope.go:117] "RemoveContainer" containerID="28c47e499d47eba2a32f24c6d1eb33a2851c69c9376562d86e1b71e662ce0bb0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.215719 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.243057 4735 scope.go:117] "RemoveContainer" containerID="eb2196fe673298fd8e33ae46b0df1a1c75009e33f8c7ed31adff4ceefea57f4d" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.274646 4735 scope.go:117] "RemoveContainer" containerID="156ac2f42f920a1a333bc28b2ae024f5f7088f40be66c08f3ffbcc665b865d60" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.286124 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.293522 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.302482 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:28:15 crc kubenswrapper[4735]: E1122 08:28:15.303143 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="ceilometer-central-agent" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.303168 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="ceilometer-central-agent" Nov 22 08:28:15 crc kubenswrapper[4735]: E1122 08:28:15.303185 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="sg-core" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.303194 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="sg-core" Nov 22 08:28:15 crc kubenswrapper[4735]: E1122 08:28:15.303216 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="ceilometer-notification-agent" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.303225 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="ceilometer-notification-agent" Nov 22 08:28:15 crc kubenswrapper[4735]: E1122 08:28:15.303242 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="proxy-httpd" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.303250 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="proxy-httpd" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.305881 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="proxy-httpd" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.305943 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="ceilometer-central-agent" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.306006 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="sg-core" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.306024 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" containerName="ceilometer-notification-agent" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.322361 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.327932 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.353170 4735 scope.go:117] "RemoveContainer" containerID="a1d7ffb7b045a61e047393f56dc98e283830cfb2ad0de2b568dd67a6d6fcd433" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.357042 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.363364 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.363439 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.386001 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.386808 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zl2z\" (UniqueName: \"kubernetes.io/projected/bb8cd234-aea0-4676-b64f-dac57de70951-kube-api-access-4zl2z\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.386869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.386914 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-log-httpd\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.387023 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-config-data\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.387081 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.387179 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-run-httpd\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.387286 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-scripts\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.489562 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zl2z\" (UniqueName: \"kubernetes.io/projected/bb8cd234-aea0-4676-b64f-dac57de70951-kube-api-access-4zl2z\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.489642 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.489688 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-log-httpd\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.489780 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-config-data\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.489829 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.489910 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-run-httpd\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.489950 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-scripts\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.489989 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.490472 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-log-httpd\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.490754 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-run-httpd\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.493892 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.494190 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.494823 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.499835 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-scripts\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.500555 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-config-data\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.507774 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zl2z\" (UniqueName: \"kubernetes.io/projected/bb8cd234-aea0-4676-b64f-dac57de70951-kube-api-access-4zl2z\") pod \"ceilometer-0\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " pod="openstack/ceilometer-0" Nov 22 08:28:15 crc kubenswrapper[4735]: I1122 08:28:15.708006 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:28:16 crc kubenswrapper[4735]: I1122 08:28:16.131831 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:28:16 crc kubenswrapper[4735]: I1122 08:28:16.132164 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:28:16 crc kubenswrapper[4735]: I1122 08:28:16.132210 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:28:16 crc kubenswrapper[4735]: I1122 08:28:16.133078 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:28:16 crc kubenswrapper[4735]: I1122 08:28:16.133140 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" gracePeriod=600 Nov 22 08:28:16 crc kubenswrapper[4735]: I1122 08:28:16.194837 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:28:16 crc kubenswrapper[4735]: I1122 08:28:16.229310 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerStarted","Data":"1cd35d3dc223d131bdc08890715e85b6eb38b84c14a01689059d62c45f2d46b8"} Nov 22 08:28:16 crc kubenswrapper[4735]: E1122 08:28:16.262065 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:28:17 crc kubenswrapper[4735]: I1122 08:28:17.245751 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerStarted","Data":"a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d"} Nov 22 08:28:17 crc kubenswrapper[4735]: I1122 08:28:17.248598 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" exitCode=0 Nov 22 08:28:17 crc kubenswrapper[4735]: I1122 08:28:17.248649 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7"} Nov 22 08:28:17 crc kubenswrapper[4735]: I1122 08:28:17.248726 4735 scope.go:117] "RemoveContainer" containerID="4afcbebd71a886d2b368e2d15dd469e8f84d1fd0d397678cdfa5ebd38afdfac1" Nov 22 08:28:17 crc kubenswrapper[4735]: I1122 08:28:17.249490 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:28:17 crc kubenswrapper[4735]: E1122 08:28:17.249907 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:28:17 crc kubenswrapper[4735]: I1122 08:28:17.277414 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89192c39-d1e5-4ca8-a1f2-b4c976ae7802" path="/var/lib/kubelet/pods/89192c39-d1e5-4ca8-a1f2-b4c976ae7802/volumes" Nov 22 08:28:18 crc kubenswrapper[4735]: I1122 08:28:18.267867 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerStarted","Data":"9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f"} Nov 22 08:28:18 crc kubenswrapper[4735]: I1122 08:28:18.268401 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerStarted","Data":"40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9"} Nov 22 08:28:20 crc kubenswrapper[4735]: I1122 08:28:20.295254 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerStarted","Data":"d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd"} Nov 22 08:28:20 crc kubenswrapper[4735]: I1122 08:28:20.296226 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:28:20 crc kubenswrapper[4735]: I1122 08:28:20.332562 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.225140062 podStartE2EDuration="5.332527202s" podCreationTimestamp="2025-11-22 08:28:15 +0000 UTC" firstStartedPulling="2025-11-22 08:28:16.198649435 +0000 UTC m=+1517.802988040" lastFinishedPulling="2025-11-22 08:28:19.306036575 +0000 UTC m=+1520.910375180" observedRunningTime="2025-11-22 08:28:20.322914285 +0000 UTC m=+1521.927252890" watchObservedRunningTime="2025-11-22 08:28:20.332527202 +0000 UTC m=+1521.936865807" Nov 22 08:28:20 crc kubenswrapper[4735]: I1122 08:28:20.644659 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 22 08:28:29 crc kubenswrapper[4735]: I1122 08:28:29.290079 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:28:29 crc kubenswrapper[4735]: E1122 08:28:29.291065 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:28:44 crc kubenswrapper[4735]: I1122 08:28:44.263943 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:28:44 crc kubenswrapper[4735]: E1122 08:28:44.265910 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:28:45 crc kubenswrapper[4735]: I1122 08:28:45.719678 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 08:28:56 crc kubenswrapper[4735]: I1122 08:28:56.263506 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:28:56 crc kubenswrapper[4735]: E1122 08:28:56.264576 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.423836 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-kvl85"] Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.435079 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-kvl85"] Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.541599 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-7wl2v"] Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.544159 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.560953 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-7wl2v"] Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.654733 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-combined-ca-bundle\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.654796 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqs6n\" (UniqueName: \"kubernetes.io/projected/ca27f7d6-a068-47bb-a286-1941df09fdec-kube-api-access-fqs6n\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.654853 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-config-data\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.756381 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-combined-ca-bundle\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.756444 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqs6n\" (UniqueName: \"kubernetes.io/projected/ca27f7d6-a068-47bb-a286-1941df09fdec-kube-api-access-fqs6n\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.756543 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-config-data\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.763306 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-config-data\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.763383 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-combined-ca-bundle\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.778707 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqs6n\" (UniqueName: \"kubernetes.io/projected/ca27f7d6-a068-47bb-a286-1941df09fdec-kube-api-access-fqs6n\") pod \"heat-db-sync-7wl2v\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:58 crc kubenswrapper[4735]: I1122 08:28:58.906516 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7wl2v" Nov 22 08:28:59 crc kubenswrapper[4735]: I1122 08:28:59.281390 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73149717-cef1-4092-b8af-5cceb0cd4830" path="/var/lib/kubelet/pods/73149717-cef1-4092-b8af-5cceb0cd4830/volumes" Nov 22 08:28:59 crc kubenswrapper[4735]: I1122 08:28:59.455791 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-7wl2v"] Nov 22 08:28:59 crc kubenswrapper[4735]: I1122 08:28:59.458177 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:28:59 crc kubenswrapper[4735]: I1122 08:28:59.800729 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7wl2v" event={"ID":"ca27f7d6-a068-47bb-a286-1941df09fdec","Type":"ContainerStarted","Data":"f6ef4aea06d5a65218f5cb17e0d0af8808a6e3ef76c78f5c284de9933c6e1d3e"} Nov 22 08:29:00 crc kubenswrapper[4735]: I1122 08:29:00.675444 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:29:00 crc kubenswrapper[4735]: I1122 08:29:00.676442 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="ceilometer-central-agent" containerID="cri-o://a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d" gracePeriod=30 Nov 22 08:29:00 crc kubenswrapper[4735]: I1122 08:29:00.676935 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="proxy-httpd" containerID="cri-o://d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd" gracePeriod=30 Nov 22 08:29:00 crc kubenswrapper[4735]: I1122 08:29:00.676982 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="sg-core" containerID="cri-o://9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f" gracePeriod=30 Nov 22 08:29:00 crc kubenswrapper[4735]: I1122 08:29:00.677014 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="ceilometer-notification-agent" containerID="cri-o://40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9" gracePeriod=30 Nov 22 08:29:00 crc kubenswrapper[4735]: I1122 08:29:00.712224 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:29:00 crc kubenswrapper[4735]: I1122 08:29:00.854068 4735 generic.go:334] "Generic (PLEG): container finished" podID="bb8cd234-aea0-4676-b64f-dac57de70951" containerID="9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f" exitCode=2 Nov 22 08:29:00 crc kubenswrapper[4735]: I1122 08:29:00.854144 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerDied","Data":"9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f"} Nov 22 08:29:01 crc kubenswrapper[4735]: I1122 08:29:01.872768 4735 generic.go:334] "Generic (PLEG): container finished" podID="bb8cd234-aea0-4676-b64f-dac57de70951" containerID="d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd" exitCode=0 Nov 22 08:29:01 crc kubenswrapper[4735]: I1122 08:29:01.873084 4735 generic.go:334] "Generic (PLEG): container finished" podID="bb8cd234-aea0-4676-b64f-dac57de70951" containerID="a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d" exitCode=0 Nov 22 08:29:01 crc kubenswrapper[4735]: I1122 08:29:01.872849 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerDied","Data":"d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd"} Nov 22 08:29:01 crc kubenswrapper[4735]: I1122 08:29:01.873129 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerDied","Data":"a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d"} Nov 22 08:29:02 crc kubenswrapper[4735]: I1122 08:29:02.403953 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:29:03 crc kubenswrapper[4735]: I1122 08:29:03.601837 4735 scope.go:117] "RemoveContainer" containerID="ce33ea0327b60ba451a8c4b8f72bad7f53c1270c3e4e6bb4012f9b3c0c94f30c" Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.365246 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerName="rabbitmq" containerID="cri-o://c844c326444f1cb2205eb64a4c985689b5a47f324d64e8e485fb7134dbd49654" gracePeriod=604796 Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.843510 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.943422 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-sg-core-conf-yaml\") pod \"bb8cd234-aea0-4676-b64f-dac57de70951\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.943720 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zl2z\" (UniqueName: \"kubernetes.io/projected/bb8cd234-aea0-4676-b64f-dac57de70951-kube-api-access-4zl2z\") pod \"bb8cd234-aea0-4676-b64f-dac57de70951\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.943920 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-combined-ca-bundle\") pod \"bb8cd234-aea0-4676-b64f-dac57de70951\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.943975 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-log-httpd\") pod \"bb8cd234-aea0-4676-b64f-dac57de70951\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.944004 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-scripts\") pod \"bb8cd234-aea0-4676-b64f-dac57de70951\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.944045 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-ceilometer-tls-certs\") pod \"bb8cd234-aea0-4676-b64f-dac57de70951\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.944060 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-config-data\") pod \"bb8cd234-aea0-4676-b64f-dac57de70951\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.944099 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-run-httpd\") pod \"bb8cd234-aea0-4676-b64f-dac57de70951\" (UID: \"bb8cd234-aea0-4676-b64f-dac57de70951\") " Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.945119 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bb8cd234-aea0-4676-b64f-dac57de70951" (UID: "bb8cd234-aea0-4676-b64f-dac57de70951"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.946014 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bb8cd234-aea0-4676-b64f-dac57de70951" (UID: "bb8cd234-aea0-4676-b64f-dac57de70951"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.950556 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb8cd234-aea0-4676-b64f-dac57de70951-kube-api-access-4zl2z" (OuterVolumeSpecName: "kube-api-access-4zl2z") pod "bb8cd234-aea0-4676-b64f-dac57de70951" (UID: "bb8cd234-aea0-4676-b64f-dac57de70951"). InnerVolumeSpecName "kube-api-access-4zl2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.955525 4735 generic.go:334] "Generic (PLEG): container finished" podID="bb8cd234-aea0-4676-b64f-dac57de70951" containerID="40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9" exitCode=0 Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.955579 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerDied","Data":"40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9"} Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.955611 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb8cd234-aea0-4676-b64f-dac57de70951","Type":"ContainerDied","Data":"1cd35d3dc223d131bdc08890715e85b6eb38b84c14a01689059d62c45f2d46b8"} Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.955631 4735 scope.go:117] "RemoveContainer" containerID="d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd" Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.955826 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.983756 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-scripts" (OuterVolumeSpecName: "scripts") pod "bb8cd234-aea0-4676-b64f-dac57de70951" (UID: "bb8cd234-aea0-4676-b64f-dac57de70951"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:05 crc kubenswrapper[4735]: I1122 08:29:05.983851 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bb8cd234-aea0-4676-b64f-dac57de70951" (UID: "bb8cd234-aea0-4676-b64f-dac57de70951"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.029992 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bb8cd234-aea0-4676-b64f-dac57de70951" (UID: "bb8cd234-aea0-4676-b64f-dac57de70951"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.046690 4735 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.046727 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.046743 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.046753 4735 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb8cd234-aea0-4676-b64f-dac57de70951-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.046762 4735 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.046771 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zl2z\" (UniqueName: \"kubernetes.io/projected/bb8cd234-aea0-4676-b64f-dac57de70951-kube-api-access-4zl2z\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.053058 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb8cd234-aea0-4676-b64f-dac57de70951" (UID: "bb8cd234-aea0-4676-b64f-dac57de70951"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.071702 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-config-data" (OuterVolumeSpecName: "config-data") pod "bb8cd234-aea0-4676-b64f-dac57de70951" (UID: "bb8cd234-aea0-4676-b64f-dac57de70951"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.148810 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.148846 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb8cd234-aea0-4676-b64f-dac57de70951-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.290443 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.300597 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.315721 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:29:06 crc kubenswrapper[4735]: E1122 08:29:06.316204 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="ceilometer-notification-agent" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.316223 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="ceilometer-notification-agent" Nov 22 08:29:06 crc kubenswrapper[4735]: E1122 08:29:06.316246 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="ceilometer-central-agent" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.316253 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="ceilometer-central-agent" Nov 22 08:29:06 crc kubenswrapper[4735]: E1122 08:29:06.316275 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="sg-core" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.316281 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="sg-core" Nov 22 08:29:06 crc kubenswrapper[4735]: E1122 08:29:06.316309 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="proxy-httpd" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.316316 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="proxy-httpd" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.316565 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="proxy-httpd" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.316589 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="ceilometer-notification-agent" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.316598 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="ceilometer-central-agent" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.316617 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" containerName="sg-core" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.321237 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.325033 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.325150 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.325219 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.327792 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.352841 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.352901 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-scripts\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.352946 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.353003 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-run-httpd\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.353024 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-config-data\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.353076 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnbjt\" (UniqueName: \"kubernetes.io/projected/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-kube-api-access-qnbjt\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.353102 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.353361 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-log-httpd\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.455755 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.455822 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-scripts\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.455849 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.455916 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-run-httpd\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.455946 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-config-data\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.456005 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnbjt\" (UniqueName: \"kubernetes.io/projected/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-kube-api-access-qnbjt\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.456033 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.456107 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-log-httpd\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.456707 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-log-httpd\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.457744 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-run-httpd\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.462604 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-scripts\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.463672 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.463694 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.464161 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-config-data\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.468901 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.477445 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnbjt\" (UniqueName: \"kubernetes.io/projected/be212b2c-6753-40d9-9bee-0dfe2ff10dbc-kube-api-access-qnbjt\") pod \"ceilometer-0\" (UID: \"be212b2c-6753-40d9-9bee-0dfe2ff10dbc\") " pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.644004 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:29:06 crc kubenswrapper[4735]: I1122 08:29:06.818858 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerName="rabbitmq" containerID="cri-o://c30125bdc4dbd76c5af54ee9db061d48b5996cf4d250fc32b82110f4c3dc513f" gracePeriod=604796 Nov 22 08:29:07 crc kubenswrapper[4735]: I1122 08:29:07.264096 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:29:07 crc kubenswrapper[4735]: E1122 08:29:07.264357 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:29:07 crc kubenswrapper[4735]: I1122 08:29:07.278523 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb8cd234-aea0-4676-b64f-dac57de70951" path="/var/lib/kubelet/pods/bb8cd234-aea0-4676-b64f-dac57de70951/volumes" Nov 22 08:29:07 crc kubenswrapper[4735]: I1122 08:29:07.453227 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: connect: connection refused" Nov 22 08:29:08 crc kubenswrapper[4735]: I1122 08:29:08.040007 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.133:5671: connect: connection refused" Nov 22 08:29:10 crc kubenswrapper[4735]: I1122 08:29:10.497683 4735 scope.go:117] "RemoveContainer" containerID="9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f" Nov 22 08:29:12 crc kubenswrapper[4735]: I1122 08:29:12.059191 4735 generic.go:334] "Generic (PLEG): container finished" podID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerID="c844c326444f1cb2205eb64a4c985689b5a47f324d64e8e485fb7134dbd49654" exitCode=0 Nov 22 08:29:12 crc kubenswrapper[4735]: I1122 08:29:12.059561 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10","Type":"ContainerDied","Data":"c844c326444f1cb2205eb64a4c985689b5a47f324d64e8e485fb7134dbd49654"} Nov 22 08:29:13 crc kubenswrapper[4735]: I1122 08:29:13.076290 4735 generic.go:334] "Generic (PLEG): container finished" podID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerID="c30125bdc4dbd76c5af54ee9db061d48b5996cf4d250fc32b82110f4c3dc513f" exitCode=0 Nov 22 08:29:13 crc kubenswrapper[4735]: I1122 08:29:13.076366 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5ee793a0-c437-47a3-b05d-369a1730d6b1","Type":"ContainerDied","Data":"c30125bdc4dbd76c5af54ee9db061d48b5996cf4d250fc32b82110f4c3dc513f"} Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.508114 4735 scope.go:117] "RemoveContainer" containerID="40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9" Nov 22 08:29:16 crc kubenswrapper[4735]: E1122 08:29:16.520865 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Nov 22 08:29:16 crc kubenswrapper[4735]: E1122 08:29:16.520933 4735 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Nov 22 08:29:16 crc kubenswrapper[4735]: E1122 08:29:16.521066 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fqs6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-7wl2v_openstack(ca27f7d6-a068-47bb-a286-1941df09fdec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:29:16 crc kubenswrapper[4735]: E1122 08:29:16.522559 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-7wl2v" podUID="ca27f7d6-a068-47bb-a286-1941df09fdec" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.676267 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.708750 4735 scope.go:117] "RemoveContainer" containerID="a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.720427 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-config-data\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.720673 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-plugins-conf\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.720709 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.720777 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-tls\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.720854 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-pod-info\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.720908 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-plugins\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.720950 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-erlang-cookie\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.720968 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-server-conf\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.721035 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgb2t\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-kube-api-access-lgb2t\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.721080 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-erlang-cookie-secret\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.721114 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-confd\") pod \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\" (UID: \"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10\") " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.721766 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.723527 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.726057 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.727964 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-pod-info" (OuterVolumeSpecName: "pod-info") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.728328 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.731450 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.737823 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.762815 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-kube-api-access-lgb2t" (OuterVolumeSpecName: "kube-api-access-lgb2t") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "kube-api-access-lgb2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.823836 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.823864 4735 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-pod-info\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.823873 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.823882 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.823893 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgb2t\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-kube-api-access-lgb2t\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.823901 4735 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.823909 4735 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.823936 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.833918 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-config-data" (OuterVolumeSpecName: "config-data") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.853297 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-server-conf" (OuterVolumeSpecName: "server-conf") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.901988 4735 scope.go:117] "RemoveContainer" containerID="d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd" Nov 22 08:29:16 crc kubenswrapper[4735]: E1122 08:29:16.902523 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd\": container with ID starting with d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd not found: ID does not exist" containerID="d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.902558 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd"} err="failed to get container status \"d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd\": rpc error: code = NotFound desc = could not find container \"d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd\": container with ID starting with d6c666e59f5ec68b03d239081b6e0a6ec0769ee3768bab960e8207927db083bd not found: ID does not exist" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.902587 4735 scope.go:117] "RemoveContainer" containerID="9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f" Nov 22 08:29:16 crc kubenswrapper[4735]: E1122 08:29:16.902854 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f\": container with ID starting with 9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f not found: ID does not exist" containerID="9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.902890 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f"} err="failed to get container status \"9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f\": rpc error: code = NotFound desc = could not find container \"9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f\": container with ID starting with 9666c58d3fd5c45d165e5bcdad2baa73394c51b74ce66865ff99a2d77d40856f not found: ID does not exist" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.902910 4735 scope.go:117] "RemoveContainer" containerID="40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9" Nov 22 08:29:16 crc kubenswrapper[4735]: E1122 08:29:16.903181 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9\": container with ID starting with 40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9 not found: ID does not exist" containerID="40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.903204 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9"} err="failed to get container status \"40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9\": rpc error: code = NotFound desc = could not find container \"40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9\": container with ID starting with 40b50719991f76d125707fa355d01c727d2f9086c845f73756a669efe2e64bf9 not found: ID does not exist" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.903222 4735 scope.go:117] "RemoveContainer" containerID="a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d" Nov 22 08:29:16 crc kubenswrapper[4735]: E1122 08:29:16.903670 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d\": container with ID starting with a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d not found: ID does not exist" containerID="a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.903697 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d"} err="failed to get container status \"a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d\": rpc error: code = NotFound desc = could not find container \"a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d\": container with ID starting with a8971bf6e3d5d31a56ca2f9550e28d3e83eb340e4940c5223a69619bd58e0f3d not found: ID does not exist" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.908035 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.926844 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.926878 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.926890 4735 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-server-conf\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.932341 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" (UID: "bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:16 crc kubenswrapper[4735]: I1122 08:29:16.954654 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028094 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ee793a0-c437-47a3-b05d-369a1730d6b1-erlang-cookie-secret\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028205 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-plugins\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028232 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f6c2\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-kube-api-access-4f6c2\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028279 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-server-conf\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028443 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-confd\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028479 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-tls\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028519 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ee793a0-c437-47a3-b05d-369a1730d6b1-pod-info\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028554 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028639 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-config-data\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028687 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-plugins-conf\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028745 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-erlang-cookie\") pod \"5ee793a0-c437-47a3-b05d-369a1730d6b1\" (UID: \"5ee793a0-c437-47a3-b05d-369a1730d6b1\") " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.028938 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.029375 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.029395 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.031869 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.032512 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-kube-api-access-4f6c2" (OuterVolumeSpecName: "kube-api-access-4f6c2") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "kube-api-access-4f6c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.032578 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.038024 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.039943 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5ee793a0-c437-47a3-b05d-369a1730d6b1-pod-info" (OuterVolumeSpecName: "pod-info") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.040399 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.042879 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee793a0-c437-47a3-b05d-369a1730d6b1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.122660 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-config-data" (OuterVolumeSpecName: "config-data") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.127718 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-server-conf" (OuterVolumeSpecName: "server-conf") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.134058 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.134064 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10","Type":"ContainerDied","Data":"818516c632fa956d6867999897cd1febf5bd027c197923a272842e62e8aa06ae"} Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.134701 4735 scope.go:117] "RemoveContainer" containerID="c844c326444f1cb2205eb64a4c985689b5a47f324d64e8e485fb7134dbd49654" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136097 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136138 4735 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ee793a0-c437-47a3-b05d-369a1730d6b1-pod-info\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136164 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136179 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136193 4735 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136247 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136288 4735 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ee793a0-c437-47a3-b05d-369a1730d6b1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136299 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f6c2\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-kube-api-access-4f6c2\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.136311 4735 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ee793a0-c437-47a3-b05d-369a1730d6b1-server-conf\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.145018 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.146049 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5ee793a0-c437-47a3-b05d-369a1730d6b1","Type":"ContainerDied","Data":"f01069603ce7dd2091a4a771d1c6acc21d979dd103325c8234adabcb49a87dc8"} Nov 22 08:29:17 crc kubenswrapper[4735]: E1122 08:29:17.153184 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-7wl2v" podUID="ca27f7d6-a068-47bb-a286-1941df09fdec" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.171683 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.190068 4735 scope.go:117] "RemoveContainer" containerID="639b9585adb4796e0e70cf58fd13241ab15ccbd2b150f54e3680a54b17b7030f" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.231626 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.233654 4735 scope.go:117] "RemoveContainer" containerID="c30125bdc4dbd76c5af54ee9db061d48b5996cf4d250fc32b82110f4c3dc513f" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.238899 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.258469 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.281649 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5ee793a0-c437-47a3-b05d-369a1730d6b1" (UID: "5ee793a0-c437-47a3-b05d-369a1730d6b1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.286823 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" path="/var/lib/kubelet/pods/bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10/volumes" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.290487 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: E1122 08:29:17.290873 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerName="rabbitmq" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.290889 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerName="rabbitmq" Nov 22 08:29:17 crc kubenswrapper[4735]: E1122 08:29:17.290910 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerName="setup-container" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.290916 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerName="setup-container" Nov 22 08:29:17 crc kubenswrapper[4735]: E1122 08:29:17.290956 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerName="setup-container" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.290963 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerName="setup-container" Nov 22 08:29:17 crc kubenswrapper[4735]: E1122 08:29:17.290970 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerName="rabbitmq" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.290976 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerName="rabbitmq" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.293392 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" containerName="rabbitmq" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.293444 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde10ac7-b2b9-4a6d-a2c8-4453f6e37f10" containerName="rabbitmq" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.296072 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.304926 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.305175 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.304979 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mk25c" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.305017 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.305855 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.305044 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.305141 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.306946 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.327514 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.340725 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-config-data\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.340769 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.340793 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.340840 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.340869 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ac149374-df2c-4533-b55e-071b860e88af-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.340889 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.340936 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlrs5\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-kube-api-access-wlrs5\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.340966 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.341007 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.341043 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ac149374-df2c-4533-b55e-071b860e88af-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.341110 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.341190 4735 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ee793a0-c437-47a3-b05d-369a1730d6b1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.371147 4735 scope.go:117] "RemoveContainer" containerID="b6b68c78ecac956d2a32b20d00cbfd5998f52fb623cb15919609639792d34e37" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.442946 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.443216 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ac149374-df2c-4533-b55e-071b860e88af-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.443328 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.443442 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlrs5\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-kube-api-access-wlrs5\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.443645 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.443748 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.443863 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ac149374-df2c-4533-b55e-071b860e88af-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.444010 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.443492 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.444165 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.444375 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-config-data\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.444208 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.444527 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.446877 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.445216 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.445673 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-config-data\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.445042 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ac149374-df2c-4533-b55e-071b860e88af-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.449109 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ac149374-df2c-4533-b55e-071b860e88af-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.449350 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.453146 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ac149374-df2c-4533-b55e-071b860e88af-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.453478 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.462493 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlrs5\" (UniqueName: \"kubernetes.io/projected/ac149374-df2c-4533-b55e-071b860e88af-kube-api-access-wlrs5\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.507988 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"ac149374-df2c-4533-b55e-071b860e88af\") " pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.578401 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.590985 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.603693 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.605664 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.611808 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.612225 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.612654 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.613024 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.613439 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.613923 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.614377 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-j67jk" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.615141 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.651972 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652064 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652093 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/019d2c19-2804-4eb6-a150-8dc1163f28c6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652109 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652158 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652180 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/019d2c19-2804-4eb6-a150-8dc1163f28c6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652246 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vptjq\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-kube-api-access-vptjq\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652265 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652307 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652327 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.652349 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.670603 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754559 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/019d2c19-2804-4eb6-a150-8dc1163f28c6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754620 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754715 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754745 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/019d2c19-2804-4eb6-a150-8dc1163f28c6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vptjq\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-kube-api-access-vptjq\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754874 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754937 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754939 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.754967 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.755454 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.755566 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.755766 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.756686 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.756746 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.757309 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.757647 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.758600 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/019d2c19-2804-4eb6-a150-8dc1163f28c6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.760311 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/019d2c19-2804-4eb6-a150-8dc1163f28c6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.760316 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/019d2c19-2804-4eb6-a150-8dc1163f28c6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.765834 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.766335 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.776301 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vptjq\" (UniqueName: \"kubernetes.io/projected/019d2c19-2804-4eb6-a150-8dc1163f28c6-kube-api-access-vptjq\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.861793 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"019d2c19-2804-4eb6-a150-8dc1163f28c6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:17 crc kubenswrapper[4735]: I1122 08:29:17.945737 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.158928 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.170764 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be212b2c-6753-40d9-9bee-0dfe2ff10dbc","Type":"ContainerStarted","Data":"ec960c7c0b12983ce8b21a3a3538c323b90a2f4e3f5ae8292f18eecd96b00a4c"} Nov 22 08:29:18 crc kubenswrapper[4735]: W1122 08:29:18.228640 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac149374_df2c_4533_b55e_071b860e88af.slice/crio-809cebdeba4a6f58eaf84ec8b97c04c6d4bcdce2f6d4432ef7833b783629264d WatchSource:0}: Error finding container 809cebdeba4a6f58eaf84ec8b97c04c6d4bcdce2f6d4432ef7833b783629264d: Status 404 returned error can't find the container with id 809cebdeba4a6f58eaf84ec8b97c04c6d4bcdce2f6d4432ef7833b783629264d Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.326056 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.849431 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-7fngp"] Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.853349 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.866662 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.893732 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-7fngp"] Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.991324 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.991381 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.991439 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.991496 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdqq4\" (UniqueName: \"kubernetes.io/projected/9bf6d124-e3b5-4e48-b444-303b0404fb15-kube-api-access-vdqq4\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.991526 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-config\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.991567 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:18 crc kubenswrapper[4735]: I1122 08:29:18.991725 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.019380 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-7fngp"] Nov 22 08:29:19 crc kubenswrapper[4735]: E1122 08:29:19.023269 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-vdqq4 openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" podUID="9bf6d124-e3b5-4e48-b444-303b0404fb15" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.060086 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-8x2j8"] Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.062004 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.089054 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-8x2j8"] Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.094899 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.095128 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.095278 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.095309 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.095406 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.095451 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdqq4\" (UniqueName: \"kubernetes.io/projected/9bf6d124-e3b5-4e48-b444-303b0404fb15-kube-api-access-vdqq4\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.095505 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-config\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.096250 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.114288 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-config\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.119206 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.125307 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.133939 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdqq4\" (UniqueName: \"kubernetes.io/projected/9bf6d124-e3b5-4e48-b444-303b0404fb15-kube-api-access-vdqq4\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.137368 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.149313 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-7fngp\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.197721 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.197801 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.198292 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.198354 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-config\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.198453 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.198707 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.198821 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gxd5\" (UniqueName: \"kubernetes.io/projected/61b5bedf-84ff-4392-9435-0f53415cc123-kube-api-access-4gxd5\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.200604 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"019d2c19-2804-4eb6-a150-8dc1163f28c6","Type":"ContainerStarted","Data":"b9888e872a187a5f49639a779ba4675d31ef91d248dc2d59092931dc19a78f4a"} Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.202672 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ac149374-df2c-4533-b55e-071b860e88af","Type":"ContainerStarted","Data":"809cebdeba4a6f58eaf84ec8b97c04c6d4bcdce2f6d4432ef7833b783629264d"} Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.202729 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.282902 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:29:19 crc kubenswrapper[4735]: E1122 08:29:19.283256 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.283871 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ee793a0-c437-47a3-b05d-369a1730d6b1" path="/var/lib/kubelet/pods/5ee793a0-c437-47a3-b05d-369a1730d6b1/volumes" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.289524 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.302177 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.302252 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.302316 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.302368 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-config\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.302424 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.302614 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.302757 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gxd5\" (UniqueName: \"kubernetes.io/projected/61b5bedf-84ff-4392-9435-0f53415cc123-kube-api-access-4gxd5\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.303920 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.304444 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.306370 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.306508 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.306974 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.307282 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61b5bedf-84ff-4392-9435-0f53415cc123-config\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.330815 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gxd5\" (UniqueName: \"kubernetes.io/projected/61b5bedf-84ff-4392-9435-0f53415cc123-kube-api-access-4gxd5\") pod \"dnsmasq-dns-6f6df4f56c-8x2j8\" (UID: \"61b5bedf-84ff-4392-9435-0f53415cc123\") " pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.407790 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-sb\") pod \"9bf6d124-e3b5-4e48-b444-303b0404fb15\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.408036 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-swift-storage-0\") pod \"9bf6d124-e3b5-4e48-b444-303b0404fb15\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.408102 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-nb\") pod \"9bf6d124-e3b5-4e48-b444-303b0404fb15\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.408129 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-config\") pod \"9bf6d124-e3b5-4e48-b444-303b0404fb15\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.408251 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdqq4\" (UniqueName: \"kubernetes.io/projected/9bf6d124-e3b5-4e48-b444-303b0404fb15-kube-api-access-vdqq4\") pod \"9bf6d124-e3b5-4e48-b444-303b0404fb15\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.408283 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-svc\") pod \"9bf6d124-e3b5-4e48-b444-303b0404fb15\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.408340 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-openstack-edpm-ipam\") pod \"9bf6d124-e3b5-4e48-b444-303b0404fb15\" (UID: \"9bf6d124-e3b5-4e48-b444-303b0404fb15\") " Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.410968 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9bf6d124-e3b5-4e48-b444-303b0404fb15" (UID: "9bf6d124-e3b5-4e48-b444-303b0404fb15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.411088 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "9bf6d124-e3b5-4e48-b444-303b0404fb15" (UID: "9bf6d124-e3b5-4e48-b444-303b0404fb15"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.411311 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9bf6d124-e3b5-4e48-b444-303b0404fb15" (UID: "9bf6d124-e3b5-4e48-b444-303b0404fb15"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.411383 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9bf6d124-e3b5-4e48-b444-303b0404fb15" (UID: "9bf6d124-e3b5-4e48-b444-303b0404fb15"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.411394 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9bf6d124-e3b5-4e48-b444-303b0404fb15" (UID: "9bf6d124-e3b5-4e48-b444-303b0404fb15"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.411936 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.411947 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-config" (OuterVolumeSpecName: "config") pod "9bf6d124-e3b5-4e48-b444-303b0404fb15" (UID: "9bf6d124-e3b5-4e48-b444-303b0404fb15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.469970 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bf6d124-e3b5-4e48-b444-303b0404fb15-kube-api-access-vdqq4" (OuterVolumeSpecName: "kube-api-access-vdqq4") pod "9bf6d124-e3b5-4e48-b444-303b0404fb15" (UID: "9bf6d124-e3b5-4e48-b444-303b0404fb15"). InnerVolumeSpecName "kube-api-access-vdqq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.509903 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.509939 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.509951 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdqq4\" (UniqueName: \"kubernetes.io/projected/9bf6d124-e3b5-4e48-b444-303b0404fb15-kube-api-access-vdqq4\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.509960 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.509989 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.510001 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:19 crc kubenswrapper[4735]: I1122 08:29:19.510009 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bf6d124-e3b5-4e48-b444-303b0404fb15-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:20 crc kubenswrapper[4735]: I1122 08:29:20.130057 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-8x2j8"] Nov 22 08:29:20 crc kubenswrapper[4735]: I1122 08:29:20.214700 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-7fngp" Nov 22 08:29:20 crc kubenswrapper[4735]: I1122 08:29:20.277750 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-7fngp"] Nov 22 08:29:20 crc kubenswrapper[4735]: I1122 08:29:20.289953 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-7fngp"] Nov 22 08:29:21 crc kubenswrapper[4735]: I1122 08:29:21.230730 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"019d2c19-2804-4eb6-a150-8dc1163f28c6","Type":"ContainerStarted","Data":"46f397823a60a8cc71d76838960e5691d8124bbe3b3f46494040cacc59bbd1b4"} Nov 22 08:29:21 crc kubenswrapper[4735]: I1122 08:29:21.245827 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ac149374-df2c-4533-b55e-071b860e88af","Type":"ContainerStarted","Data":"eb6bbf320526f7b75a7e2de7c443b9914aa855dff4ee68cc2f810b123d35e5a2"} Nov 22 08:29:21 crc kubenswrapper[4735]: I1122 08:29:21.294283 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bf6d124-e3b5-4e48-b444-303b0404fb15" path="/var/lib/kubelet/pods/9bf6d124-e3b5-4e48-b444-303b0404fb15/volumes" Nov 22 08:29:21 crc kubenswrapper[4735]: W1122 08:29:21.828708 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61b5bedf_84ff_4392_9435_0f53415cc123.slice/crio-972a1405d8073f97cbf0373af4df803a223d678ac53cce1bfb83b41087eec307 WatchSource:0}: Error finding container 972a1405d8073f97cbf0373af4df803a223d678ac53cce1bfb83b41087eec307: Status 404 returned error can't find the container with id 972a1405d8073f97cbf0373af4df803a223d678ac53cce1bfb83b41087eec307 Nov 22 08:29:22 crc kubenswrapper[4735]: I1122 08:29:22.264391 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" event={"ID":"61b5bedf-84ff-4392-9435-0f53415cc123","Type":"ContainerStarted","Data":"972a1405d8073f97cbf0373af4df803a223d678ac53cce1bfb83b41087eec307"} Nov 22 08:29:23 crc kubenswrapper[4735]: I1122 08:29:23.280122 4735 generic.go:334] "Generic (PLEG): container finished" podID="61b5bedf-84ff-4392-9435-0f53415cc123" containerID="97f81fc745bcd4d66194bfdd1e2484597fae7ce9bcc68153fc8b37815e7d3f7a" exitCode=0 Nov 22 08:29:23 crc kubenswrapper[4735]: I1122 08:29:23.280248 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" event={"ID":"61b5bedf-84ff-4392-9435-0f53415cc123","Type":"ContainerDied","Data":"97f81fc745bcd4d66194bfdd1e2484597fae7ce9bcc68153fc8b37815e7d3f7a"} Nov 22 08:29:23 crc kubenswrapper[4735]: I1122 08:29:23.283738 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be212b2c-6753-40d9-9bee-0dfe2ff10dbc","Type":"ContainerStarted","Data":"2cd8850b9a887f91c7f239286c95a2a1cc2ea3bb89b7601680fc230fb308a932"} Nov 22 08:29:24 crc kubenswrapper[4735]: I1122 08:29:24.299396 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" event={"ID":"61b5bedf-84ff-4392-9435-0f53415cc123","Type":"ContainerStarted","Data":"4cb5a8cf52ef8b397181b67d5a0e51baa342ca124bcbb3cf21d90b72c948cd01"} Nov 22 08:29:24 crc kubenswrapper[4735]: I1122 08:29:24.300192 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:24 crc kubenswrapper[4735]: I1122 08:29:24.302542 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be212b2c-6753-40d9-9bee-0dfe2ff10dbc","Type":"ContainerStarted","Data":"fe2230ae6a6fbbc244a2b4f0600e350f1e857611a31911a8f09a882bf90f8616"} Nov 22 08:29:24 crc kubenswrapper[4735]: I1122 08:29:24.325136 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" podStartSLOduration=5.325105096 podStartE2EDuration="5.325105096s" podCreationTimestamp="2025-11-22 08:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:29:24.323034009 +0000 UTC m=+1585.927372634" watchObservedRunningTime="2025-11-22 08:29:24.325105096 +0000 UTC m=+1585.929443701" Nov 22 08:29:25 crc kubenswrapper[4735]: I1122 08:29:25.316676 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be212b2c-6753-40d9-9bee-0dfe2ff10dbc","Type":"ContainerStarted","Data":"2d1f853dc8c19f8a8cc9ad9c56121502a920107f7032317a4e543c01a4cf0b35"} Nov 22 08:29:26 crc kubenswrapper[4735]: I1122 08:29:26.330191 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be212b2c-6753-40d9-9bee-0dfe2ff10dbc","Type":"ContainerStarted","Data":"cac32178190482b9493eff97a7f1b0c146d0ea3529d3f1045715b146c3307262"} Nov 22 08:29:26 crc kubenswrapper[4735]: I1122 08:29:26.330812 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:29:26 crc kubenswrapper[4735]: I1122 08:29:26.353524 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=12.079853353 podStartE2EDuration="20.353502889s" podCreationTimestamp="2025-11-22 08:29:06 +0000 UTC" firstStartedPulling="2025-11-22 08:29:17.296800456 +0000 UTC m=+1578.901139061" lastFinishedPulling="2025-11-22 08:29:25.570449992 +0000 UTC m=+1587.174788597" observedRunningTime="2025-11-22 08:29:26.347604836 +0000 UTC m=+1587.951943461" watchObservedRunningTime="2025-11-22 08:29:26.353502889 +0000 UTC m=+1587.957841494" Nov 22 08:29:29 crc kubenswrapper[4735]: I1122 08:29:29.414927 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-8x2j8" Nov 22 08:29:29 crc kubenswrapper[4735]: I1122 08:29:29.509907 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh"] Nov 22 08:29:29 crc kubenswrapper[4735]: I1122 08:29:29.510749 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" podUID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" containerName="dnsmasq-dns" containerID="cri-o://9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851" gracePeriod=10 Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.104433 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.196018 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-config\") pod \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.196158 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-nb\") pod \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.196209 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc\") pod \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.196327 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtc6s\" (UniqueName: \"kubernetes.io/projected/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-kube-api-access-wtc6s\") pod \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.196395 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-sb\") pod \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.196525 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-swift-storage-0\") pod \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.204896 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-kube-api-access-wtc6s" (OuterVolumeSpecName: "kube-api-access-wtc6s") pod "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" (UID: "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb"). InnerVolumeSpecName "kube-api-access-wtc6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.263064 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" (UID: "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.263841 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:29:30 crc kubenswrapper[4735]: E1122 08:29:30.264385 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.277502 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" (UID: "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.300027 4735 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.300054 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtc6s\" (UniqueName: \"kubernetes.io/projected/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-kube-api-access-wtc6s\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.300063 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.300744 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-config" (OuterVolumeSpecName: "config") pod "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" (UID: "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:30 crc kubenswrapper[4735]: E1122 08:29:30.306599 4735 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc podName:d3e2a652-4bac-4667-b3ea-e20cdc04d9cb nodeName:}" failed. No retries permitted until 2025-11-22 08:29:30.806566557 +0000 UTC m=+1592.410905162 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc") pod "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" (UID: "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb") : error deleting /var/lib/kubelet/pods/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb/volume-subpaths: remove /var/lib/kubelet/pods/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb/volume-subpaths: no such file or directory Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.307346 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" (UID: "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.379911 4735 generic.go:334] "Generic (PLEG): container finished" podID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" containerID="9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851" exitCode=0 Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.379951 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" event={"ID":"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb","Type":"ContainerDied","Data":"9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851"} Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.379976 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" event={"ID":"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb","Type":"ContainerDied","Data":"873066076a75a58f6233f3bcbf80bfa222d501f0a6013e0e1390eb9877d3884a"} Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.379991 4735 scope.go:117] "RemoveContainer" containerID="9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.380106 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.403130 4735 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.403164 4735 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.410941 4735 scope.go:117] "RemoveContainer" containerID="b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.445449 4735 scope.go:117] "RemoveContainer" containerID="9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851" Nov 22 08:29:30 crc kubenswrapper[4735]: E1122 08:29:30.445824 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851\": container with ID starting with 9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851 not found: ID does not exist" containerID="9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.445872 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851"} err="failed to get container status \"9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851\": rpc error: code = NotFound desc = could not find container \"9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851\": container with ID starting with 9749cc78b12e03f5023fe5a7c90cac219d101a12c645ad2a529da63fcde96851 not found: ID does not exist" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.445897 4735 scope.go:117] "RemoveContainer" containerID="b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703" Nov 22 08:29:30 crc kubenswrapper[4735]: E1122 08:29:30.446333 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703\": container with ID starting with b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703 not found: ID does not exist" containerID="b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.446358 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703"} err="failed to get container status \"b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703\": rpc error: code = NotFound desc = could not find container \"b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703\": container with ID starting with b1dde797c6a82e76fc21a9beb18ae9ac4f515e7ee871efc109bf304f339fa703 not found: ID does not exist" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.811863 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc\") pod \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\" (UID: \"d3e2a652-4bac-4667-b3ea-e20cdc04d9cb\") " Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.813003 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" (UID: "d3e2a652-4bac-4667-b3ea-e20cdc04d9cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:29:30 crc kubenswrapper[4735]: I1122 08:29:30.914430 4735 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:31 crc kubenswrapper[4735]: I1122 08:29:31.031616 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh"] Nov 22 08:29:31 crc kubenswrapper[4735]: I1122 08:29:31.052234 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-cxzfh"] Nov 22 08:29:31 crc kubenswrapper[4735]: I1122 08:29:31.278761 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" path="/var/lib/kubelet/pods/d3e2a652-4bac-4667-b3ea-e20cdc04d9cb/volumes" Nov 22 08:29:33 crc kubenswrapper[4735]: I1122 08:29:33.421912 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7wl2v" event={"ID":"ca27f7d6-a068-47bb-a286-1941df09fdec","Type":"ContainerStarted","Data":"3d590a4899629b1574a875232b61d00da1a8bcceb36ce04219789036ce50496f"} Nov 22 08:29:33 crc kubenswrapper[4735]: I1122 08:29:33.451712 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-7wl2v" podStartSLOduration=2.353718263 podStartE2EDuration="35.45168067s" podCreationTimestamp="2025-11-22 08:28:58 +0000 UTC" firstStartedPulling="2025-11-22 08:28:59.45794597 +0000 UTC m=+1561.062284575" lastFinishedPulling="2025-11-22 08:29:32.555908377 +0000 UTC m=+1594.160246982" observedRunningTime="2025-11-22 08:29:33.44482582 +0000 UTC m=+1595.049164475" watchObservedRunningTime="2025-11-22 08:29:33.45168067 +0000 UTC m=+1595.056019305" Nov 22 08:29:36 crc kubenswrapper[4735]: I1122 08:29:36.471215 4735 generic.go:334] "Generic (PLEG): container finished" podID="ca27f7d6-a068-47bb-a286-1941df09fdec" containerID="3d590a4899629b1574a875232b61d00da1a8bcceb36ce04219789036ce50496f" exitCode=0 Nov 22 08:29:36 crc kubenswrapper[4735]: I1122 08:29:36.471297 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7wl2v" event={"ID":"ca27f7d6-a068-47bb-a286-1941df09fdec","Type":"ContainerDied","Data":"3d590a4899629b1574a875232b61d00da1a8bcceb36ce04219789036ce50496f"} Nov 22 08:29:36 crc kubenswrapper[4735]: I1122 08:29:36.657283 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 08:29:37 crc kubenswrapper[4735]: I1122 08:29:37.911265 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7wl2v" Nov 22 08:29:37 crc kubenswrapper[4735]: I1122 08:29:37.996246 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-config-data\") pod \"ca27f7d6-a068-47bb-a286-1941df09fdec\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " Nov 22 08:29:37 crc kubenswrapper[4735]: I1122 08:29:37.996519 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-combined-ca-bundle\") pod \"ca27f7d6-a068-47bb-a286-1941df09fdec\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " Nov 22 08:29:37 crc kubenswrapper[4735]: I1122 08:29:37.996560 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqs6n\" (UniqueName: \"kubernetes.io/projected/ca27f7d6-a068-47bb-a286-1941df09fdec-kube-api-access-fqs6n\") pod \"ca27f7d6-a068-47bb-a286-1941df09fdec\" (UID: \"ca27f7d6-a068-47bb-a286-1941df09fdec\") " Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.003041 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca27f7d6-a068-47bb-a286-1941df09fdec-kube-api-access-fqs6n" (OuterVolumeSpecName: "kube-api-access-fqs6n") pod "ca27f7d6-a068-47bb-a286-1941df09fdec" (UID: "ca27f7d6-a068-47bb-a286-1941df09fdec"). InnerVolumeSpecName "kube-api-access-fqs6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.029280 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca27f7d6-a068-47bb-a286-1941df09fdec" (UID: "ca27f7d6-a068-47bb-a286-1941df09fdec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.099038 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.099079 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqs6n\" (UniqueName: \"kubernetes.io/projected/ca27f7d6-a068-47bb-a286-1941df09fdec-kube-api-access-fqs6n\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.111815 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-config-data" (OuterVolumeSpecName: "config-data") pod "ca27f7d6-a068-47bb-a286-1941df09fdec" (UID: "ca27f7d6-a068-47bb-a286-1941df09fdec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.200958 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca27f7d6-a068-47bb-a286-1941df09fdec-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.497277 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7wl2v" event={"ID":"ca27f7d6-a068-47bb-a286-1941df09fdec","Type":"ContainerDied","Data":"f6ef4aea06d5a65218f5cb17e0d0af8808a6e3ef76c78f5c284de9933c6e1d3e"} Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.497316 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6ef4aea06d5a65218f5cb17e0d0af8808a6e3ef76c78f5c284de9933c6e1d3e" Nov 22 08:29:38 crc kubenswrapper[4735]: I1122 08:29:38.497661 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7wl2v" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.389831 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7fc4dcbc59-gp5ct"] Nov 22 08:29:39 crc kubenswrapper[4735]: E1122 08:29:39.390709 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca27f7d6-a068-47bb-a286-1941df09fdec" containerName="heat-db-sync" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.390728 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca27f7d6-a068-47bb-a286-1941df09fdec" containerName="heat-db-sync" Nov 22 08:29:39 crc kubenswrapper[4735]: E1122 08:29:39.390742 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" containerName="init" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.390751 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" containerName="init" Nov 22 08:29:39 crc kubenswrapper[4735]: E1122 08:29:39.390785 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" containerName="dnsmasq-dns" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.390793 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" containerName="dnsmasq-dns" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.391115 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca27f7d6-a068-47bb-a286-1941df09fdec" containerName="heat-db-sync" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.391143 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e2a652-4bac-4667-b3ea-e20cdc04d9cb" containerName="dnsmasq-dns" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.392158 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.424731 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7fc4dcbc59-gp5ct"] Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.480153 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-68bf475768-f4hbn"] Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.482215 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.516538 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5d98d484bf-d5s85"] Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.518497 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531490 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-combined-ca-bundle\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531527 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-config-data-custom\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531560 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhp8g\" (UniqueName: \"kubernetes.io/projected/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-kube-api-access-mhp8g\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531579 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvtkt\" (UniqueName: \"kubernetes.io/projected/6a03b596-b9b1-4efd-b55d-711cb33f7f78-kube-api-access-xvtkt\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531597 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-config-data\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-combined-ca-bundle\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531720 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-internal-tls-certs\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531764 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-config-data-custom\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531793 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-config-data\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.531852 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-public-tls-certs\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.561505 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-68bf475768-f4hbn"] Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.634715 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhp8g\" (UniqueName: \"kubernetes.io/projected/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-kube-api-access-mhp8g\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.634767 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvtkt\" (UniqueName: \"kubernetes.io/projected/6a03b596-b9b1-4efd-b55d-711cb33f7f78-kube-api-access-xvtkt\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.634795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-config-data\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.634850 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-combined-ca-bundle\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.634938 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-config-data\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.634959 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-combined-ca-bundle\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635004 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-internal-tls-certs\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635052 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-public-tls-certs\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635086 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-config-data-custom\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635118 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2w59\" (UniqueName: \"kubernetes.io/projected/a4486150-d780-4df9-a6a4-9aa0b78c468f-kube-api-access-x2w59\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635138 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-internal-tls-certs\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635167 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-config-data\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635254 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-public-tls-certs\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635313 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-combined-ca-bundle\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635333 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-config-data-custom\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.635356 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-config-data-custom\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.649729 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-config-data-custom\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.651939 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-config-data\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.655897 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-combined-ca-bundle\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.660058 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-internal-tls-certs\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.660211 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-public-tls-certs\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.669339 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-config-data\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.670233 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-config-data-custom\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.683871 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5d98d484bf-d5s85"] Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.694259 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvtkt\" (UniqueName: \"kubernetes.io/projected/6a03b596-b9b1-4efd-b55d-711cb33f7f78-kube-api-access-xvtkt\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.700770 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a03b596-b9b1-4efd-b55d-711cb33f7f78-combined-ca-bundle\") pod \"heat-engine-7fc4dcbc59-gp5ct\" (UID: \"6a03b596-b9b1-4efd-b55d-711cb33f7f78\") " pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.708482 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhp8g\" (UniqueName: \"kubernetes.io/projected/7eaaafaf-ea3f-4ba9-b558-e128f1fdde64-kube-api-access-mhp8g\") pod \"heat-api-68bf475768-f4hbn\" (UID: \"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64\") " pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.717577 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.737272 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-config-data\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.737582 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-combined-ca-bundle\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.737779 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-public-tls-certs\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.737912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2w59\" (UniqueName: \"kubernetes.io/projected/a4486150-d780-4df9-a6a4-9aa0b78c468f-kube-api-access-x2w59\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.738021 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-internal-tls-certs\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.738277 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-config-data-custom\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.741918 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-config-data\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.743159 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-internal-tls-certs\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.744544 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-combined-ca-bundle\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.744579 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-config-data-custom\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.748072 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4486150-d780-4df9-a6a4-9aa0b78c468f-public-tls-certs\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.764218 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2w59\" (UniqueName: \"kubernetes.io/projected/a4486150-d780-4df9-a6a4-9aa0b78c468f-kube-api-access-x2w59\") pod \"heat-cfnapi-5d98d484bf-d5s85\" (UID: \"a4486150-d780-4df9-a6a4-9aa0b78c468f\") " pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.818905 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:39 crc kubenswrapper[4735]: I1122 08:29:39.852110 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:40 crc kubenswrapper[4735]: I1122 08:29:40.234512 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7fc4dcbc59-gp5ct"] Nov 22 08:29:40 crc kubenswrapper[4735]: I1122 08:29:40.440581 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-68bf475768-f4hbn"] Nov 22 08:29:40 crc kubenswrapper[4735]: I1122 08:29:40.538912 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5d98d484bf-d5s85"] Nov 22 08:29:40 crc kubenswrapper[4735]: I1122 08:29:40.571789 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7fc4dcbc59-gp5ct" event={"ID":"6a03b596-b9b1-4efd-b55d-711cb33f7f78","Type":"ContainerStarted","Data":"818efa4c2cea50bb3c5f48bd417d82ba3fc96cc1c5f2e880f1614bfd0f6e7a7c"} Nov 22 08:29:40 crc kubenswrapper[4735]: I1122 08:29:40.581960 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-68bf475768-f4hbn" event={"ID":"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64","Type":"ContainerStarted","Data":"76ea1daa3e31f8c79a5ea00340b01ebb055e63f98411639365c85033c68fc0da"} Nov 22 08:29:40 crc kubenswrapper[4735]: I1122 08:29:40.583418 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d98d484bf-d5s85" event={"ID":"a4486150-d780-4df9-a6a4-9aa0b78c468f","Type":"ContainerStarted","Data":"fec2201566e69afbdfef036e1aa62d9fb0a651ba711fac5b40d05c7d4221756f"} Nov 22 08:29:41 crc kubenswrapper[4735]: I1122 08:29:41.597188 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7fc4dcbc59-gp5ct" event={"ID":"6a03b596-b9b1-4efd-b55d-711cb33f7f78","Type":"ContainerStarted","Data":"78a628f04c8ccdfb21a7bfb837e279d2e779503e0461857e96ce4542db144131"} Nov 22 08:29:41 crc kubenswrapper[4735]: I1122 08:29:41.597429 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:41 crc kubenswrapper[4735]: I1122 08:29:41.624312 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7fc4dcbc59-gp5ct" podStartSLOduration=2.624289912 podStartE2EDuration="2.624289912s" podCreationTimestamp="2025-11-22 08:29:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:29:41.614971445 +0000 UTC m=+1603.219310090" watchObservedRunningTime="2025-11-22 08:29:41.624289912 +0000 UTC m=+1603.228628517" Nov 22 08:29:42 crc kubenswrapper[4735]: I1122 08:29:42.268282 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:29:42 crc kubenswrapper[4735]: E1122 08:29:42.268791 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:29:42 crc kubenswrapper[4735]: I1122 08:29:42.609835 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-68bf475768-f4hbn" event={"ID":"7eaaafaf-ea3f-4ba9-b558-e128f1fdde64","Type":"ContainerStarted","Data":"6816fd38071d71487180140e768d01b4f36bc6eb0f723f1af1720a302712a2e0"} Nov 22 08:29:42 crc kubenswrapper[4735]: I1122 08:29:42.609911 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:42 crc kubenswrapper[4735]: I1122 08:29:42.612185 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d98d484bf-d5s85" event={"ID":"a4486150-d780-4df9-a6a4-9aa0b78c468f","Type":"ContainerStarted","Data":"ddcbc9d204b6f00ac239051765f3ed8c0b78d7a104706f9873ddc04f5f7f5977"} Nov 22 08:29:42 crc kubenswrapper[4735]: I1122 08:29:42.638483 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-68bf475768-f4hbn" podStartSLOduration=1.871728619 podStartE2EDuration="3.638449877s" podCreationTimestamp="2025-11-22 08:29:39 +0000 UTC" firstStartedPulling="2025-11-22 08:29:40.452008697 +0000 UTC m=+1602.056347302" lastFinishedPulling="2025-11-22 08:29:42.218729955 +0000 UTC m=+1603.823068560" observedRunningTime="2025-11-22 08:29:42.636189325 +0000 UTC m=+1604.240527930" watchObservedRunningTime="2025-11-22 08:29:42.638449877 +0000 UTC m=+1604.242788482" Nov 22 08:29:42 crc kubenswrapper[4735]: I1122 08:29:42.668588 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5d98d484bf-d5s85" podStartSLOduration=2.010597757 podStartE2EDuration="3.66856613s" podCreationTimestamp="2025-11-22 08:29:39 +0000 UTC" firstStartedPulling="2025-11-22 08:29:40.542984991 +0000 UTC m=+1602.147323596" lastFinishedPulling="2025-11-22 08:29:42.200953374 +0000 UTC m=+1603.805291969" observedRunningTime="2025-11-22 08:29:42.658992545 +0000 UTC m=+1604.263331150" watchObservedRunningTime="2025-11-22 08:29:42.66856613 +0000 UTC m=+1604.272904735" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.278643 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7"] Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.279975 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.284626 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.284665 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.284719 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.284837 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.294741 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7"] Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.436192 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.436752 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.436942 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqmnh\" (UniqueName: \"kubernetes.io/projected/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-kube-api-access-pqmnh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.437143 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.539984 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.540431 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.540625 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.540763 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqmnh\" (UniqueName: \"kubernetes.io/projected/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-kube-api-access-pqmnh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.545499 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.546173 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.548860 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.566143 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqmnh\" (UniqueName: \"kubernetes.io/projected/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-kube-api-access-pqmnh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.606873 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:29:43 crc kubenswrapper[4735]: I1122 08:29:43.621881 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:44 crc kubenswrapper[4735]: I1122 08:29:44.368050 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7"] Nov 22 08:29:44 crc kubenswrapper[4735]: W1122 08:29:44.369910 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde67d031_9e4d_4ef8_bb18_f4d9fb04835c.slice/crio-b7cca9e006c1190a13b6001c9bc8c42ca9e39c7c1db571a763c0c633203d301b WatchSource:0}: Error finding container b7cca9e006c1190a13b6001c9bc8c42ca9e39c7c1db571a763c0c633203d301b: Status 404 returned error can't find the container with id b7cca9e006c1190a13b6001c9bc8c42ca9e39c7c1db571a763c0c633203d301b Nov 22 08:29:44 crc kubenswrapper[4735]: I1122 08:29:44.636742 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" event={"ID":"de67d031-9e4d-4ef8-bb18-f4d9fb04835c","Type":"ContainerStarted","Data":"b7cca9e006c1190a13b6001c9bc8c42ca9e39c7c1db571a763c0c633203d301b"} Nov 22 08:29:53 crc kubenswrapper[4735]: I1122 08:29:53.755527 4735 generic.go:334] "Generic (PLEG): container finished" podID="ac149374-df2c-4533-b55e-071b860e88af" containerID="eb6bbf320526f7b75a7e2de7c443b9914aa855dff4ee68cc2f810b123d35e5a2" exitCode=0 Nov 22 08:29:53 crc kubenswrapper[4735]: I1122 08:29:53.755677 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ac149374-df2c-4533-b55e-071b860e88af","Type":"ContainerDied","Data":"eb6bbf320526f7b75a7e2de7c443b9914aa855dff4ee68cc2f810b123d35e5a2"} Nov 22 08:29:53 crc kubenswrapper[4735]: I1122 08:29:53.760001 4735 generic.go:334] "Generic (PLEG): container finished" podID="019d2c19-2804-4eb6-a150-8dc1163f28c6" containerID="46f397823a60a8cc71d76838960e5691d8124bbe3b3f46494040cacc59bbd1b4" exitCode=0 Nov 22 08:29:53 crc kubenswrapper[4735]: I1122 08:29:53.760058 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"019d2c19-2804-4eb6-a150-8dc1163f28c6","Type":"ContainerDied","Data":"46f397823a60a8cc71d76838960e5691d8124bbe3b3f46494040cacc59bbd1b4"} Nov 22 08:29:54 crc kubenswrapper[4735]: I1122 08:29:54.263762 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:29:54 crc kubenswrapper[4735]: E1122 08:29:54.264101 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:29:54 crc kubenswrapper[4735]: I1122 08:29:54.793050 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5d98d484bf-d5s85" Nov 22 08:29:54 crc kubenswrapper[4735]: I1122 08:29:54.801920 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-68bf475768-f4hbn" Nov 22 08:29:54 crc kubenswrapper[4735]: I1122 08:29:54.940486 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7d54dc7bc4-cfqcw"] Nov 22 08:29:54 crc kubenswrapper[4735]: I1122 08:29:54.940709 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" podUID="06d273c9-b1e2-46bc-a6e3-26b26d24e966" containerName="heat-cfnapi" containerID="cri-o://c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2" gracePeriod=60 Nov 22 08:29:54 crc kubenswrapper[4735]: I1122 08:29:54.961891 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-b64b95447-gmvn2"] Nov 22 08:29:54 crc kubenswrapper[4735]: I1122 08:29:54.962099 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-b64b95447-gmvn2" podUID="5d438585-a36f-4e86-ab10-df80ac1f8340" containerName="heat-api" containerID="cri-o://6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873" gracePeriod=60 Nov 22 08:29:55 crc kubenswrapper[4735]: I1122 08:29:55.786689 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" event={"ID":"de67d031-9e4d-4ef8-bb18-f4d9fb04835c","Type":"ContainerStarted","Data":"7efe7beef545f0671da8e749fac0e3e0894b0ef1870db1a1ee358ba06203aecd"} Nov 22 08:29:55 crc kubenswrapper[4735]: I1122 08:29:55.789609 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ac149374-df2c-4533-b55e-071b860e88af","Type":"ContainerStarted","Data":"002bf4a3432a1f0b6e1d32f42f584038e060d4b0fec15d173b81c2a834f8eb97"} Nov 22 08:29:55 crc kubenswrapper[4735]: I1122 08:29:55.792294 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"019d2c19-2804-4eb6-a150-8dc1163f28c6","Type":"ContainerStarted","Data":"e6b72c692f14369ff09491e22e9a1940695699c3e09ac4838c92f8402404e2f4"} Nov 22 08:29:55 crc kubenswrapper[4735]: I1122 08:29:55.792515 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:29:55 crc kubenswrapper[4735]: I1122 08:29:55.805805 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" podStartSLOduration=2.208303158 podStartE2EDuration="12.805783014s" podCreationTimestamp="2025-11-22 08:29:43 +0000 UTC" firstStartedPulling="2025-11-22 08:29:44.372218936 +0000 UTC m=+1605.976557541" lastFinishedPulling="2025-11-22 08:29:54.969698792 +0000 UTC m=+1616.574037397" observedRunningTime="2025-11-22 08:29:55.805240229 +0000 UTC m=+1617.409578834" watchObservedRunningTime="2025-11-22 08:29:55.805783014 +0000 UTC m=+1617.410121619" Nov 22 08:29:55 crc kubenswrapper[4735]: I1122 08:29:55.831796 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.831778203 podStartE2EDuration="38.831778203s" podCreationTimestamp="2025-11-22 08:29:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:29:55.827165615 +0000 UTC m=+1617.431504210" watchObservedRunningTime="2025-11-22 08:29:55.831778203 +0000 UTC m=+1617.436116808" Nov 22 08:29:55 crc kubenswrapper[4735]: I1122 08:29:55.859275 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.859251742 podStartE2EDuration="38.859251742s" podCreationTimestamp="2025-11-22 08:29:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:29:55.847640861 +0000 UTC m=+1617.451979466" watchObservedRunningTime="2025-11-22 08:29:55.859251742 +0000 UTC m=+1617.463590347" Nov 22 08:29:57 crc kubenswrapper[4735]: I1122 08:29:57.670944 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.157200 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-b64b95447-gmvn2" podUID="5d438585-a36f-4e86-ab10-df80ac1f8340" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.224:8004/healthcheck\": read tcp 10.217.0.2:50668->10.217.0.224:8004: read: connection reset by peer" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.172820 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" podUID="06d273c9-b1e2-46bc-a6e3-26b26d24e966" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.225:8000/healthcheck\": read tcp 10.217.0.2:35738->10.217.0.225:8000: read: connection reset by peer" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.732483 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.740594 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.852352 4735 generic.go:334] "Generic (PLEG): container finished" podID="06d273c9-b1e2-46bc-a6e3-26b26d24e966" containerID="c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2" exitCode=0 Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.852406 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.852477 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" event={"ID":"06d273c9-b1e2-46bc-a6e3-26b26d24e966","Type":"ContainerDied","Data":"c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2"} Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.852527 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d54dc7bc4-cfqcw" event={"ID":"06d273c9-b1e2-46bc-a6e3-26b26d24e966","Type":"ContainerDied","Data":"219edf1a7a052034cee00fc7c095c0ddf6bbdc98efda0649760cb027b91ac394"} Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.852551 4735 scope.go:117] "RemoveContainer" containerID="c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854389 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nq88\" (UniqueName: \"kubernetes.io/projected/5d438585-a36f-4e86-ab10-df80ac1f8340-kube-api-access-4nq88\") pod \"5d438585-a36f-4e86-ab10-df80ac1f8340\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854445 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-internal-tls-certs\") pod \"5d438585-a36f-4e86-ab10-df80ac1f8340\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854491 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-combined-ca-bundle\") pod \"5d438585-a36f-4e86-ab10-df80ac1f8340\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854572 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-internal-tls-certs\") pod \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854608 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data\") pod \"5d438585-a36f-4e86-ab10-df80ac1f8340\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854700 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data\") pod \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854766 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data-custom\") pod \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854782 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-combined-ca-bundle\") pod \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854805 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data-custom\") pod \"5d438585-a36f-4e86-ab10-df80ac1f8340\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854831 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-public-tls-certs\") pod \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854925 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6jfh\" (UniqueName: \"kubernetes.io/projected/06d273c9-b1e2-46bc-a6e3-26b26d24e966-kube-api-access-r6jfh\") pod \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\" (UID: \"06d273c9-b1e2-46bc-a6e3-26b26d24e966\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.854957 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-public-tls-certs\") pod \"5d438585-a36f-4e86-ab10-df80ac1f8340\" (UID: \"5d438585-a36f-4e86-ab10-df80ac1f8340\") " Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.861715 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d438585-a36f-4e86-ab10-df80ac1f8340-kube-api-access-4nq88" (OuterVolumeSpecName: "kube-api-access-4nq88") pod "5d438585-a36f-4e86-ab10-df80ac1f8340" (UID: "5d438585-a36f-4e86-ab10-df80ac1f8340"). InnerVolumeSpecName "kube-api-access-4nq88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.863516 4735 generic.go:334] "Generic (PLEG): container finished" podID="5d438585-a36f-4e86-ab10-df80ac1f8340" containerID="6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873" exitCode=0 Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.863665 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-b64b95447-gmvn2" event={"ID":"5d438585-a36f-4e86-ab10-df80ac1f8340","Type":"ContainerDied","Data":"6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873"} Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.863701 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-b64b95447-gmvn2" event={"ID":"5d438585-a36f-4e86-ab10-df80ac1f8340","Type":"ContainerDied","Data":"81ce07b78d21ab6d845110ef21de502a0577062d88ca4a37e61feeee912ac6e9"} Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.863737 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "06d273c9-b1e2-46bc-a6e3-26b26d24e966" (UID: "06d273c9-b1e2-46bc-a6e3-26b26d24e966"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.863766 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-b64b95447-gmvn2" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.864112 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06d273c9-b1e2-46bc-a6e3-26b26d24e966-kube-api-access-r6jfh" (OuterVolumeSpecName: "kube-api-access-r6jfh") pod "06d273c9-b1e2-46bc-a6e3-26b26d24e966" (UID: "06d273c9-b1e2-46bc-a6e3-26b26d24e966"). InnerVolumeSpecName "kube-api-access-r6jfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.879833 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5d438585-a36f-4e86-ab10-df80ac1f8340" (UID: "5d438585-a36f-4e86-ab10-df80ac1f8340"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.880808 4735 scope.go:117] "RemoveContainer" containerID="c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2" Nov 22 08:29:58 crc kubenswrapper[4735]: E1122 08:29:58.881391 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2\": container with ID starting with c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2 not found: ID does not exist" containerID="c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.881421 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2"} err="failed to get container status \"c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2\": rpc error: code = NotFound desc = could not find container \"c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2\": container with ID starting with c841de9dfc1572f1de7e745bc40ab16f29a8d711e1d8b289d6ae48740616e0c2 not found: ID does not exist" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.881443 4735 scope.go:117] "RemoveContainer" containerID="6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.901761 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d438585-a36f-4e86-ab10-df80ac1f8340" (UID: "5d438585-a36f-4e86-ab10-df80ac1f8340"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.923069 4735 scope.go:117] "RemoveContainer" containerID="6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873" Nov 22 08:29:58 crc kubenswrapper[4735]: E1122 08:29:58.923719 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873\": container with ID starting with 6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873 not found: ID does not exist" containerID="6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.923783 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873"} err="failed to get container status \"6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873\": rpc error: code = NotFound desc = could not find container \"6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873\": container with ID starting with 6e8fb8f3af22e4bc0502c501552bdb43cc653c09332681e6eb0dae14aa843873 not found: ID does not exist" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.925691 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06d273c9-b1e2-46bc-a6e3-26b26d24e966" (UID: "06d273c9-b1e2-46bc-a6e3-26b26d24e966"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.942751 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5d438585-a36f-4e86-ab10-df80ac1f8340" (UID: "5d438585-a36f-4e86-ab10-df80ac1f8340"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.945009 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "06d273c9-b1e2-46bc-a6e3-26b26d24e966" (UID: "06d273c9-b1e2-46bc-a6e3-26b26d24e966"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.945866 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data" (OuterVolumeSpecName: "config-data") pod "06d273c9-b1e2-46bc-a6e3-26b26d24e966" (UID: "06d273c9-b1e2-46bc-a6e3-26b26d24e966"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.952030 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5d438585-a36f-4e86-ab10-df80ac1f8340" (UID: "5d438585-a36f-4e86-ab10-df80ac1f8340"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957814 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6jfh\" (UniqueName: \"kubernetes.io/projected/06d273c9-b1e2-46bc-a6e3-26b26d24e966-kube-api-access-r6jfh\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957847 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957857 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nq88\" (UniqueName: \"kubernetes.io/projected/5d438585-a36f-4e86-ab10-df80ac1f8340-kube-api-access-4nq88\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957867 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957877 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957888 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957904 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957915 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957926 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.957937 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.966077 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "06d273c9-b1e2-46bc-a6e3-26b26d24e966" (UID: "06d273c9-b1e2-46bc-a6e3-26b26d24e966"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:58 crc kubenswrapper[4735]: I1122 08:29:58.966667 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data" (OuterVolumeSpecName: "config-data") pod "5d438585-a36f-4e86-ab10-df80ac1f8340" (UID: "5d438585-a36f-4e86-ab10-df80ac1f8340"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.060848 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06d273c9-b1e2-46bc-a6e3-26b26d24e966-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.060888 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d438585-a36f-4e86-ab10-df80ac1f8340-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.213101 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7d54dc7bc4-cfqcw"] Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.227613 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7d54dc7bc4-cfqcw"] Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.238505 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-b64b95447-gmvn2"] Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.252114 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-b64b95447-gmvn2"] Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.285953 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06d273c9-b1e2-46bc-a6e3-26b26d24e966" path="/var/lib/kubelet/pods/06d273c9-b1e2-46bc-a6e3-26b26d24e966/volumes" Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.286693 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d438585-a36f-4e86-ab10-df80ac1f8340" path="/var/lib/kubelet/pods/5d438585-a36f-4e86-ab10-df80ac1f8340/volumes" Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.753215 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7fc4dcbc59-gp5ct" Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.800010 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6796459cfd-kjmnc"] Nov 22 08:29:59 crc kubenswrapper[4735]: I1122 08:29:59.800210 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-6796459cfd-kjmnc" podUID="1db40e01-d3c8-4cc3-a571-f21049cd6616" containerName="heat-engine" containerID="cri-o://0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" gracePeriod=60 Nov 22 08:29:59 crc kubenswrapper[4735]: E1122 08:29:59.964940 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:29:59 crc kubenswrapper[4735]: E1122 08:29:59.966695 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:29:59 crc kubenswrapper[4735]: E1122 08:29:59.967882 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:29:59 crc kubenswrapper[4735]: E1122 08:29:59.967918 4735 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6796459cfd-kjmnc" podUID="1db40e01-d3c8-4cc3-a571-f21049cd6616" containerName="heat-engine" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.174938 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx"] Nov 22 08:30:00 crc kubenswrapper[4735]: E1122 08:30:00.175582 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d273c9-b1e2-46bc-a6e3-26b26d24e966" containerName="heat-cfnapi" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.175604 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d273c9-b1e2-46bc-a6e3-26b26d24e966" containerName="heat-cfnapi" Nov 22 08:30:00 crc kubenswrapper[4735]: E1122 08:30:00.175671 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d438585-a36f-4e86-ab10-df80ac1f8340" containerName="heat-api" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.175680 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d438585-a36f-4e86-ab10-df80ac1f8340" containerName="heat-api" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.175948 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d438585-a36f-4e86-ab10-df80ac1f8340" containerName="heat-api" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.175989 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d273c9-b1e2-46bc-a6e3-26b26d24e966" containerName="heat-cfnapi" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.177006 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.179200 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.179728 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.202309 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx"] Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.284525 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42w2c\" (UniqueName: \"kubernetes.io/projected/a12655bc-7394-4051-b783-4aa456551746-kube-api-access-42w2c\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.284678 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a12655bc-7394-4051-b783-4aa456551746-secret-volume\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.284743 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a12655bc-7394-4051-b783-4aa456551746-config-volume\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.386526 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a12655bc-7394-4051-b783-4aa456551746-secret-volume\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.386912 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a12655bc-7394-4051-b783-4aa456551746-config-volume\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.387147 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42w2c\" (UniqueName: \"kubernetes.io/projected/a12655bc-7394-4051-b783-4aa456551746-kube-api-access-42w2c\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.387816 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a12655bc-7394-4051-b783-4aa456551746-config-volume\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.405027 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a12655bc-7394-4051-b783-4aa456551746-secret-volume\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.408740 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42w2c\" (UniqueName: \"kubernetes.io/projected/a12655bc-7394-4051-b783-4aa456551746-kube-api-access-42w2c\") pod \"collect-profiles-29396670-92thx\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:00 crc kubenswrapper[4735]: I1122 08:30:00.519974 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:01 crc kubenswrapper[4735]: I1122 08:30:01.116586 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx"] Nov 22 08:30:01 crc kubenswrapper[4735]: I1122 08:30:01.900188 4735 generic.go:334] "Generic (PLEG): container finished" podID="a12655bc-7394-4051-b783-4aa456551746" containerID="d3c8b6caf421b8b0aa649050f8c34b94e3a86cd62c43f490d99080f106efec4c" exitCode=0 Nov 22 08:30:01 crc kubenswrapper[4735]: I1122 08:30:01.900248 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" event={"ID":"a12655bc-7394-4051-b783-4aa456551746","Type":"ContainerDied","Data":"d3c8b6caf421b8b0aa649050f8c34b94e3a86cd62c43f490d99080f106efec4c"} Nov 22 08:30:01 crc kubenswrapper[4735]: I1122 08:30:01.900508 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" event={"ID":"a12655bc-7394-4051-b783-4aa456551746","Type":"ContainerStarted","Data":"91989c668898244654b098d0124d14d98ee56d80162bbe1d767ff4eebad0a1c5"} Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.360678 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.464186 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42w2c\" (UniqueName: \"kubernetes.io/projected/a12655bc-7394-4051-b783-4aa456551746-kube-api-access-42w2c\") pod \"a12655bc-7394-4051-b783-4aa456551746\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.464723 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a12655bc-7394-4051-b783-4aa456551746-secret-volume\") pod \"a12655bc-7394-4051-b783-4aa456551746\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.464825 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a12655bc-7394-4051-b783-4aa456551746-config-volume\") pod \"a12655bc-7394-4051-b783-4aa456551746\" (UID: \"a12655bc-7394-4051-b783-4aa456551746\") " Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.466035 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a12655bc-7394-4051-b783-4aa456551746-config-volume" (OuterVolumeSpecName: "config-volume") pod "a12655bc-7394-4051-b783-4aa456551746" (UID: "a12655bc-7394-4051-b783-4aa456551746"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.488798 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a12655bc-7394-4051-b783-4aa456551746-kube-api-access-42w2c" (OuterVolumeSpecName: "kube-api-access-42w2c") pod "a12655bc-7394-4051-b783-4aa456551746" (UID: "a12655bc-7394-4051-b783-4aa456551746"). InnerVolumeSpecName "kube-api-access-42w2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.488891 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a12655bc-7394-4051-b783-4aa456551746-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a12655bc-7394-4051-b783-4aa456551746" (UID: "a12655bc-7394-4051-b783-4aa456551746"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.567043 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42w2c\" (UniqueName: \"kubernetes.io/projected/a12655bc-7394-4051-b783-4aa456551746-kube-api-access-42w2c\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.567079 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a12655bc-7394-4051-b783-4aa456551746-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.567094 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a12655bc-7394-4051-b783-4aa456551746-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.924593 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" event={"ID":"a12655bc-7394-4051-b783-4aa456551746","Type":"ContainerDied","Data":"91989c668898244654b098d0124d14d98ee56d80162bbe1d767ff4eebad0a1c5"} Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.924642 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91989c668898244654b098d0124d14d98ee56d80162bbe1d767ff4eebad0a1c5" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.924643 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx" Nov 22 08:30:03 crc kubenswrapper[4735]: I1122 08:30:03.972648 4735 scope.go:117] "RemoveContainer" containerID="ca5814fdbfd4c6c36249028f417d2a8534621e0cd7a9496af74de53682f2201c" Nov 22 08:30:04 crc kubenswrapper[4735]: I1122 08:30:04.016906 4735 scope.go:117] "RemoveContainer" containerID="37cc336eeb0e55dacf3617c3e737ce28e02bacb73aafccee97cb3bc4b4e0c6d2" Nov 22 08:30:04 crc kubenswrapper[4735]: I1122 08:30:04.093488 4735 scope.go:117] "RemoveContainer" containerID="33d4f40ef552d8b1c5dbed8c3b3d4e62283ddfc16a37f2822441144dcf24ecc5" Nov 22 08:30:04 crc kubenswrapper[4735]: I1122 08:30:04.119587 4735 scope.go:117] "RemoveContainer" containerID="b3c9fc6ea8757266331eabc4a4429d89d80fb829b60d8538119de99da812b8ae" Nov 22 08:30:04 crc kubenswrapper[4735]: I1122 08:30:04.150263 4735 scope.go:117] "RemoveContainer" containerID="83a6187f25b3c04cb4758ec8d289340f16e25e0b8c0847b4fd93fa50b6c4ca15" Nov 22 08:30:04 crc kubenswrapper[4735]: I1122 08:30:04.180033 4735 scope.go:117] "RemoveContainer" containerID="4ceba5e68a13c3d1a5af6b8a3587abf671726c5c98bf992784ed5c5fa99ea41f" Nov 22 08:30:04 crc kubenswrapper[4735]: I1122 08:30:04.211794 4735 scope.go:117] "RemoveContainer" containerID="328ee9ef102328cae7327a50f8ee34bd45ad68bf2f30298f9f3aafa182c6cb00" Nov 22 08:30:05 crc kubenswrapper[4735]: I1122 08:30:05.266753 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:30:05 crc kubenswrapper[4735]: E1122 08:30:05.267446 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:30:06 crc kubenswrapper[4735]: I1122 08:30:06.966091 4735 generic.go:334] "Generic (PLEG): container finished" podID="de67d031-9e4d-4ef8-bb18-f4d9fb04835c" containerID="7efe7beef545f0671da8e749fac0e3e0894b0ef1870db1a1ee358ba06203aecd" exitCode=0 Nov 22 08:30:06 crc kubenswrapper[4735]: I1122 08:30:06.966131 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" event={"ID":"de67d031-9e4d-4ef8-bb18-f4d9fb04835c","Type":"ContainerDied","Data":"7efe7beef545f0671da8e749fac0e3e0894b0ef1870db1a1ee358ba06203aecd"} Nov 22 08:30:07 crc kubenswrapper[4735]: I1122 08:30:07.674755 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 22 08:30:07 crc kubenswrapper[4735]: I1122 08:30:07.950692 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:30:07 crc kubenswrapper[4735]: I1122 08:30:07.958195 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-rkfmf"] Nov 22 08:30:07 crc kubenswrapper[4735]: I1122 08:30:07.987369 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-rkfmf"] Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.143269 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-74rcz"] Nov 22 08:30:08 crc kubenswrapper[4735]: E1122 08:30:08.143834 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a12655bc-7394-4051-b783-4aa456551746" containerName="collect-profiles" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.143853 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a12655bc-7394-4051-b783-4aa456551746" containerName="collect-profiles" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.144185 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a12655bc-7394-4051-b783-4aa456551746" containerName="collect-profiles" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.145022 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.148207 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.177720 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-74rcz"] Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.304013 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-combined-ca-bundle\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.304223 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbmg6\" (UniqueName: \"kubernetes.io/projected/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-kube-api-access-jbmg6\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.304292 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-config-data\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.304373 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-scripts\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.406744 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbmg6\" (UniqueName: \"kubernetes.io/projected/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-kube-api-access-jbmg6\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.406808 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-config-data\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.406845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-scripts\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.406992 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-combined-ca-bundle\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.412616 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-combined-ca-bundle\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.415800 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-scripts\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.423418 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-config-data\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.428810 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbmg6\" (UniqueName: \"kubernetes.io/projected/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-kube-api-access-jbmg6\") pod \"aodh-db-sync-74rcz\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.474183 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.724518 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.815467 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-ssh-key\") pod \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.815651 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-inventory\") pod \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.815743 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-repo-setup-combined-ca-bundle\") pod \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.815850 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqmnh\" (UniqueName: \"kubernetes.io/projected/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-kube-api-access-pqmnh\") pod \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\" (UID: \"de67d031-9e4d-4ef8-bb18-f4d9fb04835c\") " Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.826975 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "de67d031-9e4d-4ef8-bb18-f4d9fb04835c" (UID: "de67d031-9e4d-4ef8-bb18-f4d9fb04835c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.830783 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-kube-api-access-pqmnh" (OuterVolumeSpecName: "kube-api-access-pqmnh") pod "de67d031-9e4d-4ef8-bb18-f4d9fb04835c" (UID: "de67d031-9e4d-4ef8-bb18-f4d9fb04835c"). InnerVolumeSpecName "kube-api-access-pqmnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.857857 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de67d031-9e4d-4ef8-bb18-f4d9fb04835c" (UID: "de67d031-9e4d-4ef8-bb18-f4d9fb04835c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.862681 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-inventory" (OuterVolumeSpecName: "inventory") pod "de67d031-9e4d-4ef8-bb18-f4d9fb04835c" (UID: "de67d031-9e4d-4ef8-bb18-f4d9fb04835c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.921005 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqmnh\" (UniqueName: \"kubernetes.io/projected/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-kube-api-access-pqmnh\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.921125 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.921180 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.921236 4735 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67d031-9e4d-4ef8-bb18-f4d9fb04835c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.992417 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" event={"ID":"de67d031-9e4d-4ef8-bb18-f4d9fb04835c","Type":"ContainerDied","Data":"b7cca9e006c1190a13b6001c9bc8c42ca9e39c7c1db571a763c0c633203d301b"} Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.992461 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7" Nov 22 08:30:08 crc kubenswrapper[4735]: I1122 08:30:08.992502 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7cca9e006c1190a13b6001c9bc8c42ca9e39c7c1db571a763c0c633203d301b" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.090928 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx"] Nov 22 08:30:09 crc kubenswrapper[4735]: E1122 08:30:09.091422 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de67d031-9e4d-4ef8-bb18-f4d9fb04835c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.091439 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="de67d031-9e4d-4ef8-bb18-f4d9fb04835c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.091687 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="de67d031-9e4d-4ef8-bb18-f4d9fb04835c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.092495 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.101506 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.101527 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.101586 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.103092 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.137366 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx"] Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.169915 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-74rcz"] Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.228145 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.228614 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.228648 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8wnr\" (UniqueName: \"kubernetes.io/projected/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-kube-api-access-d8wnr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.277053 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4179aca1-9f3a-4cd5-a9ad-3a301a313668" path="/var/lib/kubelet/pods/4179aca1-9f3a-4cd5-a9ad-3a301a313668/volumes" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.330351 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.330716 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.330809 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8wnr\" (UniqueName: \"kubernetes.io/projected/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-kube-api-access-d8wnr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.334189 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.336125 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.346904 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8wnr\" (UniqueName: \"kubernetes.io/projected/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-kube-api-access-d8wnr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xkqtx\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: I1122 08:30:09.419112 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:09 crc kubenswrapper[4735]: E1122 08:30:09.963939 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:30:09 crc kubenswrapper[4735]: E1122 08:30:09.978273 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:30:09 crc kubenswrapper[4735]: E1122 08:30:09.982203 4735 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 22 08:30:09 crc kubenswrapper[4735]: E1122 08:30:09.982277 4735 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6796459cfd-kjmnc" podUID="1db40e01-d3c8-4cc3-a571-f21049cd6616" containerName="heat-engine" Nov 22 08:30:10 crc kubenswrapper[4735]: I1122 08:30:10.037614 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-74rcz" event={"ID":"5b3826dc-bab1-4c6d-85dc-cecf267d0c70","Type":"ContainerStarted","Data":"7b1a4bc2039d0c232eca1d472bd5bbc13cd073b79ab6ece43132c37a862a6236"} Nov 22 08:30:10 crc kubenswrapper[4735]: I1122 08:30:10.102278 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx"] Nov 22 08:30:11 crc kubenswrapper[4735]: I1122 08:30:11.054812 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" event={"ID":"dc4a5c09-e30f-4ab9-8668-52fbd81751c5","Type":"ContainerStarted","Data":"c2ab04df6268efdce8e1674a1ed32721693b53c358b843d87b9af435c1546d53"} Nov 22 08:30:11 crc kubenswrapper[4735]: I1122 08:30:11.055641 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" event={"ID":"dc4a5c09-e30f-4ab9-8668-52fbd81751c5","Type":"ContainerStarted","Data":"22d5b716f026b8c5b93c47181c06873df51356f63eb8e925c83e1d3d40662a60"} Nov 22 08:30:11 crc kubenswrapper[4735]: I1122 08:30:11.078503 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" podStartSLOduration=1.574014051 podStartE2EDuration="2.078480611s" podCreationTimestamp="2025-11-22 08:30:09 +0000 UTC" firstStartedPulling="2025-11-22 08:30:10.122839017 +0000 UTC m=+1631.727177612" lastFinishedPulling="2025-11-22 08:30:10.627305557 +0000 UTC m=+1632.231644172" observedRunningTime="2025-11-22 08:30:11.072040823 +0000 UTC m=+1632.676379448" watchObservedRunningTime="2025-11-22 08:30:11.078480611 +0000 UTC m=+1632.682819206" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.091566 4735 generic.go:334] "Generic (PLEG): container finished" podID="dc4a5c09-e30f-4ab9-8668-52fbd81751c5" containerID="c2ab04df6268efdce8e1674a1ed32721693b53c358b843d87b9af435c1546d53" exitCode=0 Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.091664 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" event={"ID":"dc4a5c09-e30f-4ab9-8668-52fbd81751c5","Type":"ContainerDied","Data":"c2ab04df6268efdce8e1674a1ed32721693b53c358b843d87b9af435c1546d53"} Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.094163 4735 generic.go:334] "Generic (PLEG): container finished" podID="1db40e01-d3c8-4cc3-a571-f21049cd6616" containerID="0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" exitCode=0 Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.094207 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6796459cfd-kjmnc" event={"ID":"1db40e01-d3c8-4cc3-a571-f21049cd6616","Type":"ContainerDied","Data":"0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d"} Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.621900 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.767385 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-combined-ca-bundle\") pod \"1db40e01-d3c8-4cc3-a571-f21049cd6616\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.767590 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrqgs\" (UniqueName: \"kubernetes.io/projected/1db40e01-d3c8-4cc3-a571-f21049cd6616-kube-api-access-jrqgs\") pod \"1db40e01-d3c8-4cc3-a571-f21049cd6616\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.767753 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data-custom\") pod \"1db40e01-d3c8-4cc3-a571-f21049cd6616\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.767935 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data\") pod \"1db40e01-d3c8-4cc3-a571-f21049cd6616\" (UID: \"1db40e01-d3c8-4cc3-a571-f21049cd6616\") " Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.772931 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1db40e01-d3c8-4cc3-a571-f21049cd6616" (UID: "1db40e01-d3c8-4cc3-a571-f21049cd6616"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.773688 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1db40e01-d3c8-4cc3-a571-f21049cd6616-kube-api-access-jrqgs" (OuterVolumeSpecName: "kube-api-access-jrqgs") pod "1db40e01-d3c8-4cc3-a571-f21049cd6616" (UID: "1db40e01-d3c8-4cc3-a571-f21049cd6616"). InnerVolumeSpecName "kube-api-access-jrqgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.802519 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1db40e01-d3c8-4cc3-a571-f21049cd6616" (UID: "1db40e01-d3c8-4cc3-a571-f21049cd6616"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.849169 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data" (OuterVolumeSpecName: "config-data") pod "1db40e01-d3c8-4cc3-a571-f21049cd6616" (UID: "1db40e01-d3c8-4cc3-a571-f21049cd6616"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.880147 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.880187 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrqgs\" (UniqueName: \"kubernetes.io/projected/1db40e01-d3c8-4cc3-a571-f21049cd6616-kube-api-access-jrqgs\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.880198 4735 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:14 crc kubenswrapper[4735]: I1122 08:30:14.880207 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db40e01-d3c8-4cc3-a571-f21049cd6616-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.106831 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-74rcz" event={"ID":"5b3826dc-bab1-4c6d-85dc-cecf267d0c70","Type":"ContainerStarted","Data":"f0a4c584a0695c3ad4cf7fb3298c1047dc462aa6264a4cf143174834fa59e71a"} Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.109816 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6796459cfd-kjmnc" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.109856 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6796459cfd-kjmnc" event={"ID":"1db40e01-d3c8-4cc3-a571-f21049cd6616","Type":"ContainerDied","Data":"7640a03a631afc4ef7a601f7a7103d87b499d0a0cc7ffad6ecfadb22991f6336"} Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.109887 4735 scope.go:117] "RemoveContainer" containerID="0c9ae697f6cd0cdc2b94ba2da7a227fd19549ce310a283dc9f2f3aefaccfbf6d" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.135232 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-74rcz" podStartSLOduration=2.000322444 podStartE2EDuration="7.135211922s" podCreationTimestamp="2025-11-22 08:30:08 +0000 UTC" firstStartedPulling="2025-11-22 08:30:09.154694096 +0000 UTC m=+1630.759032701" lastFinishedPulling="2025-11-22 08:30:14.289583544 +0000 UTC m=+1635.893922179" observedRunningTime="2025-11-22 08:30:15.131404017 +0000 UTC m=+1636.735742632" watchObservedRunningTime="2025-11-22 08:30:15.135211922 +0000 UTC m=+1636.739550527" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.161865 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6796459cfd-kjmnc"] Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.171477 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-6796459cfd-kjmnc"] Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.278838 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1db40e01-d3c8-4cc3-a571-f21049cd6616" path="/var/lib/kubelet/pods/1db40e01-d3c8-4cc3-a571-f21049cd6616/volumes" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.661379 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.800542 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-ssh-key\") pod \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.800620 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8wnr\" (UniqueName: \"kubernetes.io/projected/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-kube-api-access-d8wnr\") pod \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.800848 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-inventory\") pod \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\" (UID: \"dc4a5c09-e30f-4ab9-8668-52fbd81751c5\") " Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.817699 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-kube-api-access-d8wnr" (OuterVolumeSpecName: "kube-api-access-d8wnr") pod "dc4a5c09-e30f-4ab9-8668-52fbd81751c5" (UID: "dc4a5c09-e30f-4ab9-8668-52fbd81751c5"). InnerVolumeSpecName "kube-api-access-d8wnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.838575 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dc4a5c09-e30f-4ab9-8668-52fbd81751c5" (UID: "dc4a5c09-e30f-4ab9-8668-52fbd81751c5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.848108 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-inventory" (OuterVolumeSpecName: "inventory") pod "dc4a5c09-e30f-4ab9-8668-52fbd81751c5" (UID: "dc4a5c09-e30f-4ab9-8668-52fbd81751c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.904125 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.904169 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8wnr\" (UniqueName: \"kubernetes.io/projected/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-kube-api-access-d8wnr\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:15 crc kubenswrapper[4735]: I1122 08:30:15.904182 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc4a5c09-e30f-4ab9-8668-52fbd81751c5-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.125127 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" event={"ID":"dc4a5c09-e30f-4ab9-8668-52fbd81751c5","Type":"ContainerDied","Data":"22d5b716f026b8c5b93c47181c06873df51356f63eb8e925c83e1d3d40662a60"} Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.125170 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22d5b716f026b8c5b93c47181c06873df51356f63eb8e925c83e1d3d40662a60" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.125199 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xkqtx" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.238729 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr"] Nov 22 08:30:16 crc kubenswrapper[4735]: E1122 08:30:16.239828 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4a5c09-e30f-4ab9-8668-52fbd81751c5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.239854 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4a5c09-e30f-4ab9-8668-52fbd81751c5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:16 crc kubenswrapper[4735]: E1122 08:30:16.239882 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db40e01-d3c8-4cc3-a571-f21049cd6616" containerName="heat-engine" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.239895 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db40e01-d3c8-4cc3-a571-f21049cd6616" containerName="heat-engine" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.240216 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1db40e01-d3c8-4cc3-a571-f21049cd6616" containerName="heat-engine" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.240255 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4a5c09-e30f-4ab9-8668-52fbd81751c5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.241294 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.245872 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.246029 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.246087 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.246189 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.256158 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr"] Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.314165 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.314410 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.314500 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq48b\" (UniqueName: \"kubernetes.io/projected/d5c15b81-20c1-46c8-8f80-dcb4e7394735-kube-api-access-sq48b\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.314652 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.416280 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.417549 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.418116 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.418173 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq48b\" (UniqueName: \"kubernetes.io/projected/d5c15b81-20c1-46c8-8f80-dcb4e7394735-kube-api-access-sq48b\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.421719 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.422553 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.437100 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.444239 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq48b\" (UniqueName: \"kubernetes.io/projected/d5c15b81-20c1-46c8-8f80-dcb4e7394735-kube-api-access-sq48b\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:16 crc kubenswrapper[4735]: I1122 08:30:16.596608 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:30:17 crc kubenswrapper[4735]: I1122 08:30:17.281424 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr"] Nov 22 08:30:18 crc kubenswrapper[4735]: I1122 08:30:18.180348 4735 generic.go:334] "Generic (PLEG): container finished" podID="5b3826dc-bab1-4c6d-85dc-cecf267d0c70" containerID="f0a4c584a0695c3ad4cf7fb3298c1047dc462aa6264a4cf143174834fa59e71a" exitCode=0 Nov 22 08:30:18 crc kubenswrapper[4735]: I1122 08:30:18.180414 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-74rcz" event={"ID":"5b3826dc-bab1-4c6d-85dc-cecf267d0c70","Type":"ContainerDied","Data":"f0a4c584a0695c3ad4cf7fb3298c1047dc462aa6264a4cf143174834fa59e71a"} Nov 22 08:30:18 crc kubenswrapper[4735]: I1122 08:30:18.183250 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" event={"ID":"d5c15b81-20c1-46c8-8f80-dcb4e7394735","Type":"ContainerStarted","Data":"29ddfc01175ae0a199f348892a4cc702e75fc8ea90962960649ac664fa627fc4"} Nov 22 08:30:18 crc kubenswrapper[4735]: I1122 08:30:18.183308 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" event={"ID":"d5c15b81-20c1-46c8-8f80-dcb4e7394735","Type":"ContainerStarted","Data":"5fe47b693ba640e509b6aa9fb0daf2ac046fad5d5de7f305b95a802ac4737a64"} Nov 22 08:30:18 crc kubenswrapper[4735]: I1122 08:30:18.221888 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" podStartSLOduration=1.805189801 podStartE2EDuration="2.221867171s" podCreationTimestamp="2025-11-22 08:30:16 +0000 UTC" firstStartedPulling="2025-11-22 08:30:17.271737578 +0000 UTC m=+1638.876076203" lastFinishedPulling="2025-11-22 08:30:17.688414968 +0000 UTC m=+1639.292753573" observedRunningTime="2025-11-22 08:30:18.214635041 +0000 UTC m=+1639.818973676" watchObservedRunningTime="2025-11-22 08:30:18.221867171 +0000 UTC m=+1639.826205776" Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.282007 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:30:19 crc kubenswrapper[4735]: E1122 08:30:19.282819 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.738794 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.903840 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-combined-ca-bundle\") pod \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.903941 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-config-data\") pod \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.904039 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbmg6\" (UniqueName: \"kubernetes.io/projected/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-kube-api-access-jbmg6\") pod \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.904077 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-scripts\") pod \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\" (UID: \"5b3826dc-bab1-4c6d-85dc-cecf267d0c70\") " Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.915752 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-kube-api-access-jbmg6" (OuterVolumeSpecName: "kube-api-access-jbmg6") pod "5b3826dc-bab1-4c6d-85dc-cecf267d0c70" (UID: "5b3826dc-bab1-4c6d-85dc-cecf267d0c70"). InnerVolumeSpecName "kube-api-access-jbmg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.915803 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-scripts" (OuterVolumeSpecName: "scripts") pod "5b3826dc-bab1-4c6d-85dc-cecf267d0c70" (UID: "5b3826dc-bab1-4c6d-85dc-cecf267d0c70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.949019 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-config-data" (OuterVolumeSpecName: "config-data") pod "5b3826dc-bab1-4c6d-85dc-cecf267d0c70" (UID: "5b3826dc-bab1-4c6d-85dc-cecf267d0c70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:19 crc kubenswrapper[4735]: I1122 08:30:19.961910 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b3826dc-bab1-4c6d-85dc-cecf267d0c70" (UID: "5b3826dc-bab1-4c6d-85dc-cecf267d0c70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:20 crc kubenswrapper[4735]: I1122 08:30:20.006825 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbmg6\" (UniqueName: \"kubernetes.io/projected/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-kube-api-access-jbmg6\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:20 crc kubenswrapper[4735]: I1122 08:30:20.006865 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:20 crc kubenswrapper[4735]: I1122 08:30:20.006876 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:20 crc kubenswrapper[4735]: I1122 08:30:20.006883 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3826dc-bab1-4c6d-85dc-cecf267d0c70-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:20 crc kubenswrapper[4735]: I1122 08:30:20.284514 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-74rcz" event={"ID":"5b3826dc-bab1-4c6d-85dc-cecf267d0c70","Type":"ContainerDied","Data":"7b1a4bc2039d0c232eca1d472bd5bbc13cd073b79ab6ece43132c37a862a6236"} Nov 22 08:30:20 crc kubenswrapper[4735]: I1122 08:30:20.284562 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b1a4bc2039d0c232eca1d472bd5bbc13cd073b79ab6ece43132c37a862a6236" Nov 22 08:30:20 crc kubenswrapper[4735]: I1122 08:30:20.284668 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-74rcz" Nov 22 08:30:23 crc kubenswrapper[4735]: I1122 08:30:23.805989 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 22 08:30:23 crc kubenswrapper[4735]: I1122 08:30:23.807769 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-api" containerID="cri-o://0798dc5eda46a091c75461a2bb9584d339a13a861f521d088d8295c91440a1d4" gracePeriod=30 Nov 22 08:30:23 crc kubenswrapper[4735]: I1122 08:30:23.807849 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-notifier" containerID="cri-o://ae00901a5c9b52914ee48e511f5380bb1657eb56dee8e3af32ea8fce46538b32" gracePeriod=30 Nov 22 08:30:23 crc kubenswrapper[4735]: I1122 08:30:23.807849 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-listener" containerID="cri-o://6462b6967208bd3cca5d8300b6b254224c25e883fc9eaa70d61b234d15c25876" gracePeriod=30 Nov 22 08:30:23 crc kubenswrapper[4735]: I1122 08:30:23.807906 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-evaluator" containerID="cri-o://e8cbdd89932353d5e2975889e319e29e0af3f59f699ee85b488473c10d8fba74" gracePeriod=30 Nov 22 08:30:24 crc kubenswrapper[4735]: I1122 08:30:24.325755 4735 generic.go:334] "Generic (PLEG): container finished" podID="72122e00-8d36-467a-9fca-d33d73ef458b" containerID="e8cbdd89932353d5e2975889e319e29e0af3f59f699ee85b488473c10d8fba74" exitCode=0 Nov 22 08:30:24 crc kubenswrapper[4735]: I1122 08:30:24.326059 4735 generic.go:334] "Generic (PLEG): container finished" podID="72122e00-8d36-467a-9fca-d33d73ef458b" containerID="0798dc5eda46a091c75461a2bb9584d339a13a861f521d088d8295c91440a1d4" exitCode=0 Nov 22 08:30:24 crc kubenswrapper[4735]: I1122 08:30:24.325864 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerDied","Data":"e8cbdd89932353d5e2975889e319e29e0af3f59f699ee85b488473c10d8fba74"} Nov 22 08:30:24 crc kubenswrapper[4735]: I1122 08:30:24.326098 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerDied","Data":"0798dc5eda46a091c75461a2bb9584d339a13a861f521d088d8295c91440a1d4"} Nov 22 08:30:28 crc kubenswrapper[4735]: I1122 08:30:28.384381 4735 generic.go:334] "Generic (PLEG): container finished" podID="72122e00-8d36-467a-9fca-d33d73ef458b" containerID="6462b6967208bd3cca5d8300b6b254224c25e883fc9eaa70d61b234d15c25876" exitCode=0 Nov 22 08:30:28 crc kubenswrapper[4735]: I1122 08:30:28.384493 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerDied","Data":"6462b6967208bd3cca5d8300b6b254224c25e883fc9eaa70d61b234d15c25876"} Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.401211 4735 generic.go:334] "Generic (PLEG): container finished" podID="72122e00-8d36-467a-9fca-d33d73ef458b" containerID="ae00901a5c9b52914ee48e511f5380bb1657eb56dee8e3af32ea8fce46538b32" exitCode=0 Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.401247 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerDied","Data":"ae00901a5c9b52914ee48e511f5380bb1657eb56dee8e3af32ea8fce46538b32"} Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.401526 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"72122e00-8d36-467a-9fca-d33d73ef458b","Type":"ContainerDied","Data":"108b387a0f538fdc450cfa4822a53dfabc4a43f4e31519c5079782b4b062fc70"} Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.401541 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="108b387a0f538fdc450cfa4822a53dfabc4a43f4e31519c5079782b4b062fc70" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.447611 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.583842 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-public-tls-certs\") pod \"72122e00-8d36-467a-9fca-d33d73ef458b\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.583928 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-config-data\") pod \"72122e00-8d36-467a-9fca-d33d73ef458b\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.584189 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mdb7\" (UniqueName: \"kubernetes.io/projected/72122e00-8d36-467a-9fca-d33d73ef458b-kube-api-access-8mdb7\") pod \"72122e00-8d36-467a-9fca-d33d73ef458b\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.584268 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-internal-tls-certs\") pod \"72122e00-8d36-467a-9fca-d33d73ef458b\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.584715 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-combined-ca-bundle\") pod \"72122e00-8d36-467a-9fca-d33d73ef458b\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.584864 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-scripts\") pod \"72122e00-8d36-467a-9fca-d33d73ef458b\" (UID: \"72122e00-8d36-467a-9fca-d33d73ef458b\") " Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.589765 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72122e00-8d36-467a-9fca-d33d73ef458b-kube-api-access-8mdb7" (OuterVolumeSpecName: "kube-api-access-8mdb7") pod "72122e00-8d36-467a-9fca-d33d73ef458b" (UID: "72122e00-8d36-467a-9fca-d33d73ef458b"). InnerVolumeSpecName "kube-api-access-8mdb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.590620 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-scripts" (OuterVolumeSpecName: "scripts") pod "72122e00-8d36-467a-9fca-d33d73ef458b" (UID: "72122e00-8d36-467a-9fca-d33d73ef458b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.646684 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "72122e00-8d36-467a-9fca-d33d73ef458b" (UID: "72122e00-8d36-467a-9fca-d33d73ef458b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.687652 4735 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.687693 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mdb7\" (UniqueName: \"kubernetes.io/projected/72122e00-8d36-467a-9fca-d33d73ef458b-kube-api-access-8mdb7\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.687713 4735 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.688284 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "72122e00-8d36-467a-9fca-d33d73ef458b" (UID: "72122e00-8d36-467a-9fca-d33d73ef458b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.723828 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72122e00-8d36-467a-9fca-d33d73ef458b" (UID: "72122e00-8d36-467a-9fca-d33d73ef458b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.743823 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-config-data" (OuterVolumeSpecName: "config-data") pod "72122e00-8d36-467a-9fca-d33d73ef458b" (UID: "72122e00-8d36-467a-9fca-d33d73ef458b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.791488 4735 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.791709 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:29 crc kubenswrapper[4735]: I1122 08:30:29.791793 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72122e00-8d36-467a-9fca-d33d73ef458b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.415996 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.460611 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.477527 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.494626 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 22 08:30:30 crc kubenswrapper[4735]: E1122 08:30:30.495161 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-evaluator" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495179 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-evaluator" Nov 22 08:30:30 crc kubenswrapper[4735]: E1122 08:30:30.495201 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-listener" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495208 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-listener" Nov 22 08:30:30 crc kubenswrapper[4735]: E1122 08:30:30.495244 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-notifier" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495255 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-notifier" Nov 22 08:30:30 crc kubenswrapper[4735]: E1122 08:30:30.495272 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-api" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495278 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-api" Nov 22 08:30:30 crc kubenswrapper[4735]: E1122 08:30:30.495297 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3826dc-bab1-4c6d-85dc-cecf267d0c70" containerName="aodh-db-sync" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495303 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3826dc-bab1-4c6d-85dc-cecf267d0c70" containerName="aodh-db-sync" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495594 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-api" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495613 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-listener" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495630 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-notifier" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495649 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" containerName="aodh-evaluator" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.495660 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3826dc-bab1-4c6d-85dc-cecf267d0c70" containerName="aodh-db-sync" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.497705 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.503737 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.503756 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-bfgs7" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.503996 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.504163 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.504195 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.508103 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.613654 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-public-tls-certs\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.613985 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-scripts\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.614284 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.614558 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-config-data\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.614721 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-internal-tls-certs\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.614874 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsbkr\" (UniqueName: \"kubernetes.io/projected/d211c51b-4464-4109-9edf-3b05feddb072-kube-api-access-gsbkr\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.717724 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsbkr\" (UniqueName: \"kubernetes.io/projected/d211c51b-4464-4109-9edf-3b05feddb072-kube-api-access-gsbkr\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.717850 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-public-tls-certs\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.717987 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-scripts\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.718033 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.718093 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-config-data\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.718134 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-internal-tls-certs\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.724714 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.728623 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-config-data\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.729075 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-scripts\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.739578 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-public-tls-certs\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.739951 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d211c51b-4464-4109-9edf-3b05feddb072-internal-tls-certs\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.739963 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsbkr\" (UniqueName: \"kubernetes.io/projected/d211c51b-4464-4109-9edf-3b05feddb072-kube-api-access-gsbkr\") pod \"aodh-0\" (UID: \"d211c51b-4464-4109-9edf-3b05feddb072\") " pod="openstack/aodh-0" Nov 22 08:30:30 crc kubenswrapper[4735]: I1122 08:30:30.818926 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 22 08:30:31 crc kubenswrapper[4735]: I1122 08:30:31.275843 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72122e00-8d36-467a-9fca-d33d73ef458b" path="/var/lib/kubelet/pods/72122e00-8d36-467a-9fca-d33d73ef458b/volumes" Nov 22 08:30:31 crc kubenswrapper[4735]: I1122 08:30:31.299762 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 22 08:30:31 crc kubenswrapper[4735]: I1122 08:30:31.457715 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d211c51b-4464-4109-9edf-3b05feddb072","Type":"ContainerStarted","Data":"b0b00eec530b60a1bcc1dcd378785b4901a2177cbfd9a08f166206980d92e113"} Nov 22 08:30:32 crc kubenswrapper[4735]: I1122 08:30:32.484945 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d211c51b-4464-4109-9edf-3b05feddb072","Type":"ContainerStarted","Data":"baeca588103f0387b21da34ea16af315c61c8039202dd8188572915ca5c0179a"} Nov 22 08:30:33 crc kubenswrapper[4735]: I1122 08:30:33.504320 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d211c51b-4464-4109-9edf-3b05feddb072","Type":"ContainerStarted","Data":"44a40a49814c301f737b810a8c46476083a227cff36d6d7acc39ca9f59e624ac"} Nov 22 08:30:34 crc kubenswrapper[4735]: I1122 08:30:34.265644 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:30:34 crc kubenswrapper[4735]: E1122 08:30:34.266323 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:30:34 crc kubenswrapper[4735]: I1122 08:30:34.520736 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d211c51b-4464-4109-9edf-3b05feddb072","Type":"ContainerStarted","Data":"ec57505281b5f16e59506e001e85d45a9452b15ea7e1377dfa60e0ceb6437dd0"} Nov 22 08:30:36 crc kubenswrapper[4735]: I1122 08:30:36.548974 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d211c51b-4464-4109-9edf-3b05feddb072","Type":"ContainerStarted","Data":"c2c3a0ee19e799f7ec737f72444468c687986502ee383aa81b8fec053f0b8852"} Nov 22 08:30:36 crc kubenswrapper[4735]: I1122 08:30:36.585731 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.581907788 podStartE2EDuration="6.585708333s" podCreationTimestamp="2025-11-22 08:30:30 +0000 UTC" firstStartedPulling="2025-11-22 08:30:31.351246818 +0000 UTC m=+1652.955585423" lastFinishedPulling="2025-11-22 08:30:35.355047363 +0000 UTC m=+1656.959385968" observedRunningTime="2025-11-22 08:30:36.576067097 +0000 UTC m=+1658.180405702" watchObservedRunningTime="2025-11-22 08:30:36.585708333 +0000 UTC m=+1658.190046948" Nov 22 08:30:47 crc kubenswrapper[4735]: I1122 08:30:47.263637 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:30:47 crc kubenswrapper[4735]: E1122 08:30:47.264644 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:30:59 crc kubenswrapper[4735]: I1122 08:30:59.281936 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:30:59 crc kubenswrapper[4735]: E1122 08:30:59.282955 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.481010 4735 scope.go:117] "RemoveContainer" containerID="8f626e28b82b7a010070b3c46d6987197699f306a73d9f8768a256b969462313" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.520360 4735 scope.go:117] "RemoveContainer" containerID="1a09894c33583101467d2d5dba0247e78fc51e16cbb109ee8485a72316353cea" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.545310 4735 scope.go:117] "RemoveContainer" containerID="348156957a5b1fee9c882df5966b47e562545ff35846ef74e2929e920c2c4add" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.584981 4735 scope.go:117] "RemoveContainer" containerID="e8b5d8776713807502178db7368884a6b93fe4c2c627a0b5fedec2752f76bab6" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.635094 4735 scope.go:117] "RemoveContainer" containerID="134a3fe6dbe39edd37b72f6318fd561fbf642c5e35fbae6c588f3735179d9dcc" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.660190 4735 scope.go:117] "RemoveContainer" containerID="ecbdc6d9ee393531116e462cba48b131422b3d535366727877f5f339eae70cf2" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.708158 4735 scope.go:117] "RemoveContainer" containerID="4a44b5d58cdf5feb94a5580fb9e450c1e9bc2b9d16727c60783a8d272f3470a4" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.763512 4735 scope.go:117] "RemoveContainer" containerID="5ce6a172cada9e142900e150f3988b104d641127aa700ee591498007b95ae4d1" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.808339 4735 scope.go:117] "RemoveContainer" containerID="e7cf03d5986fae3bfbad4db3f9f60e2e05d51d4e99ca71ecc60f0a00104622ea" Nov 22 08:31:04 crc kubenswrapper[4735]: I1122 08:31:04.845953 4735 scope.go:117] "RemoveContainer" containerID="86e1161727ab2cd73b072fdf519043d6472e246c487e974f9831da74a28761ca" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.151976 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-82tw9"] Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.154996 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.185258 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-82tw9"] Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.261962 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfd26\" (UniqueName: \"kubernetes.io/projected/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-kube-api-access-kfd26\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.262072 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-utilities\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.262089 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-catalog-content\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.363969 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfd26\" (UniqueName: \"kubernetes.io/projected/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-kube-api-access-kfd26\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.364128 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-utilities\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.364152 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-catalog-content\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.364998 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-utilities\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.365069 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-catalog-content\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.384619 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfd26\" (UniqueName: \"kubernetes.io/projected/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-kube-api-access-kfd26\") pod \"redhat-operators-82tw9\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.485071 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:06 crc kubenswrapper[4735]: I1122 08:31:06.973382 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-82tw9"] Nov 22 08:31:06 crc kubenswrapper[4735]: W1122 08:31:06.973538 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76304aea_23c2_43ca_a7f9_4a524f4c1bb3.slice/crio-dc5dd491a2fd068059c7df3de6ad3cad5eb8baac9b7196eb36ed35d07297f9a5 WatchSource:0}: Error finding container dc5dd491a2fd068059c7df3de6ad3cad5eb8baac9b7196eb36ed35d07297f9a5: Status 404 returned error can't find the container with id dc5dd491a2fd068059c7df3de6ad3cad5eb8baac9b7196eb36ed35d07297f9a5 Nov 22 08:31:07 crc kubenswrapper[4735]: I1122 08:31:07.019116 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82tw9" event={"ID":"76304aea-23c2-43ca-a7f9-4a524f4c1bb3","Type":"ContainerStarted","Data":"dc5dd491a2fd068059c7df3de6ad3cad5eb8baac9b7196eb36ed35d07297f9a5"} Nov 22 08:31:08 crc kubenswrapper[4735]: I1122 08:31:08.031313 4735 generic.go:334] "Generic (PLEG): container finished" podID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerID="7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b" exitCode=0 Nov 22 08:31:08 crc kubenswrapper[4735]: I1122 08:31:08.031382 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82tw9" event={"ID":"76304aea-23c2-43ca-a7f9-4a524f4c1bb3","Type":"ContainerDied","Data":"7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b"} Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.049761 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82tw9" event={"ID":"76304aea-23c2-43ca-a7f9-4a524f4c1bb3","Type":"ContainerStarted","Data":"49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf"} Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.546129 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-phwts"] Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.548743 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.583169 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-phwts"] Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.658613 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4krl\" (UniqueName: \"kubernetes.io/projected/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-kube-api-access-b4krl\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.658841 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-utilities\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.658890 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-catalog-content\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.760815 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-catalog-content\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.760949 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4krl\" (UniqueName: \"kubernetes.io/projected/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-kube-api-access-b4krl\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.761113 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-utilities\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.761508 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-catalog-content\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.761824 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-utilities\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.786290 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4krl\" (UniqueName: \"kubernetes.io/projected/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-kube-api-access-b4krl\") pod \"community-operators-phwts\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:09 crc kubenswrapper[4735]: I1122 08:31:09.875915 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:10 crc kubenswrapper[4735]: I1122 08:31:10.263757 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:31:10 crc kubenswrapper[4735]: E1122 08:31:10.264379 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:31:10 crc kubenswrapper[4735]: I1122 08:31:10.519393 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-phwts"] Nov 22 08:31:10 crc kubenswrapper[4735]: W1122 08:31:10.524322 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddee29d1_53a1_47f5_9a8b_89c2b3f60bb5.slice/crio-0b9c55908ce62b41a302d5a3c5ceaad5cb8ee47c688c4f7de4f08df5ee5fc3d4 WatchSource:0}: Error finding container 0b9c55908ce62b41a302d5a3c5ceaad5cb8ee47c688c4f7de4f08df5ee5fc3d4: Status 404 returned error can't find the container with id 0b9c55908ce62b41a302d5a3c5ceaad5cb8ee47c688c4f7de4f08df5ee5fc3d4 Nov 22 08:31:11 crc kubenswrapper[4735]: I1122 08:31:11.072944 4735 generic.go:334] "Generic (PLEG): container finished" podID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerID="ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc" exitCode=0 Nov 22 08:31:11 crc kubenswrapper[4735]: I1122 08:31:11.073244 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-phwts" event={"ID":"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5","Type":"ContainerDied","Data":"ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc"} Nov 22 08:31:11 crc kubenswrapper[4735]: I1122 08:31:11.073273 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-phwts" event={"ID":"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5","Type":"ContainerStarted","Data":"0b9c55908ce62b41a302d5a3c5ceaad5cb8ee47c688c4f7de4f08df5ee5fc3d4"} Nov 22 08:31:12 crc kubenswrapper[4735]: I1122 08:31:12.104947 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-phwts" event={"ID":"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5","Type":"ContainerStarted","Data":"9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed"} Nov 22 08:31:15 crc kubenswrapper[4735]: I1122 08:31:15.145105 4735 generic.go:334] "Generic (PLEG): container finished" podID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerID="9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed" exitCode=0 Nov 22 08:31:15 crc kubenswrapper[4735]: I1122 08:31:15.146544 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-phwts" event={"ID":"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5","Type":"ContainerDied","Data":"9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed"} Nov 22 08:31:15 crc kubenswrapper[4735]: I1122 08:31:15.153162 4735 generic.go:334] "Generic (PLEG): container finished" podID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerID="49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf" exitCode=0 Nov 22 08:31:15 crc kubenswrapper[4735]: I1122 08:31:15.153224 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82tw9" event={"ID":"76304aea-23c2-43ca-a7f9-4a524f4c1bb3","Type":"ContainerDied","Data":"49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf"} Nov 22 08:31:16 crc kubenswrapper[4735]: I1122 08:31:16.165646 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-phwts" event={"ID":"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5","Type":"ContainerStarted","Data":"cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78"} Nov 22 08:31:16 crc kubenswrapper[4735]: I1122 08:31:16.168952 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82tw9" event={"ID":"76304aea-23c2-43ca-a7f9-4a524f4c1bb3","Type":"ContainerStarted","Data":"ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c"} Nov 22 08:31:16 crc kubenswrapper[4735]: I1122 08:31:16.185596 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-phwts" podStartSLOduration=2.70677645 podStartE2EDuration="7.18557881s" podCreationTimestamp="2025-11-22 08:31:09 +0000 UTC" firstStartedPulling="2025-11-22 08:31:11.075403936 +0000 UTC m=+1692.679742541" lastFinishedPulling="2025-11-22 08:31:15.554206296 +0000 UTC m=+1697.158544901" observedRunningTime="2025-11-22 08:31:16.183834662 +0000 UTC m=+1697.788173267" watchObservedRunningTime="2025-11-22 08:31:16.18557881 +0000 UTC m=+1697.789917415" Nov 22 08:31:16 crc kubenswrapper[4735]: I1122 08:31:16.220901 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-82tw9" podStartSLOduration=2.68211158 podStartE2EDuration="10.220886738s" podCreationTimestamp="2025-11-22 08:31:06 +0000 UTC" firstStartedPulling="2025-11-22 08:31:08.033300832 +0000 UTC m=+1689.637639437" lastFinishedPulling="2025-11-22 08:31:15.57207599 +0000 UTC m=+1697.176414595" observedRunningTime="2025-11-22 08:31:16.211632321 +0000 UTC m=+1697.815970946" watchObservedRunningTime="2025-11-22 08:31:16.220886738 +0000 UTC m=+1697.825225343" Nov 22 08:31:16 crc kubenswrapper[4735]: I1122 08:31:16.485350 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:16 crc kubenswrapper[4735]: I1122 08:31:16.485509 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:17 crc kubenswrapper[4735]: I1122 08:31:17.556826 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82tw9" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="registry-server" probeResult="failure" output=< Nov 22 08:31:17 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:31:17 crc kubenswrapper[4735]: > Nov 22 08:31:19 crc kubenswrapper[4735]: I1122 08:31:19.876488 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:19 crc kubenswrapper[4735]: I1122 08:31:19.876980 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:20 crc kubenswrapper[4735]: I1122 08:31:20.932022 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-phwts" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="registry-server" probeResult="failure" output=< Nov 22 08:31:20 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:31:20 crc kubenswrapper[4735]: > Nov 22 08:31:24 crc kubenswrapper[4735]: I1122 08:31:24.264484 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:31:24 crc kubenswrapper[4735]: E1122 08:31:24.270544 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:31:27 crc kubenswrapper[4735]: I1122 08:31:27.545723 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82tw9" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="registry-server" probeResult="failure" output=< Nov 22 08:31:27 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:31:27 crc kubenswrapper[4735]: > Nov 22 08:31:29 crc kubenswrapper[4735]: I1122 08:31:29.959074 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:30 crc kubenswrapper[4735]: I1122 08:31:30.018003 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:30 crc kubenswrapper[4735]: I1122 08:31:30.210722 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-phwts"] Nov 22 08:31:31 crc kubenswrapper[4735]: I1122 08:31:31.356029 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-phwts" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="registry-server" containerID="cri-o://cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78" gracePeriod=2 Nov 22 08:31:31 crc kubenswrapper[4735]: I1122 08:31:31.882268 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.006016 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-utilities\") pod \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.006097 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-catalog-content\") pod \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.006345 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4krl\" (UniqueName: \"kubernetes.io/projected/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-kube-api-access-b4krl\") pod \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\" (UID: \"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5\") " Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.006782 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-utilities" (OuterVolumeSpecName: "utilities") pod "ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" (UID: "ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.007706 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.017178 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-kube-api-access-b4krl" (OuterVolumeSpecName: "kube-api-access-b4krl") pod "ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" (UID: "ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5"). InnerVolumeSpecName "kube-api-access-b4krl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.054521 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" (UID: "ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.110447 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4krl\" (UniqueName: \"kubernetes.io/projected/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-kube-api-access-b4krl\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.110504 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.378991 4735 generic.go:334] "Generic (PLEG): container finished" podID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerID="cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78" exitCode=0 Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.379061 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-phwts" event={"ID":"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5","Type":"ContainerDied","Data":"cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78"} Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.379092 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-phwts" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.379123 4735 scope.go:117] "RemoveContainer" containerID="cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.379104 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-phwts" event={"ID":"ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5","Type":"ContainerDied","Data":"0b9c55908ce62b41a302d5a3c5ceaad5cb8ee47c688c4f7de4f08df5ee5fc3d4"} Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.409291 4735 scope.go:117] "RemoveContainer" containerID="9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.423499 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-phwts"] Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.450118 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-phwts"] Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.462542 4735 scope.go:117] "RemoveContainer" containerID="ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.497570 4735 scope.go:117] "RemoveContainer" containerID="cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78" Nov 22 08:31:32 crc kubenswrapper[4735]: E1122 08:31:32.498194 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78\": container with ID starting with cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78 not found: ID does not exist" containerID="cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.498252 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78"} err="failed to get container status \"cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78\": rpc error: code = NotFound desc = could not find container \"cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78\": container with ID starting with cc3215251e67c389a70ff394d0bb9331b716a4c4c76dfa345d321ecf680e8b78 not found: ID does not exist" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.498290 4735 scope.go:117] "RemoveContainer" containerID="9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed" Nov 22 08:31:32 crc kubenswrapper[4735]: E1122 08:31:32.498707 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed\": container with ID starting with 9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed not found: ID does not exist" containerID="9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.498746 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed"} err="failed to get container status \"9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed\": rpc error: code = NotFound desc = could not find container \"9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed\": container with ID starting with 9c76cd8f193812bb674adb7ca2c9283ba7e4367cd976c8f5581ab30393b4c3ed not found: ID does not exist" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.498766 4735 scope.go:117] "RemoveContainer" containerID="ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc" Nov 22 08:31:32 crc kubenswrapper[4735]: E1122 08:31:32.499082 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc\": container with ID starting with ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc not found: ID does not exist" containerID="ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc" Nov 22 08:31:32 crc kubenswrapper[4735]: I1122 08:31:32.499122 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc"} err="failed to get container status \"ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc\": rpc error: code = NotFound desc = could not find container \"ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc\": container with ID starting with ff6ff314817fd53a0d26aff033a972732f987eefdbdf50508ac9cccc0cf65ecc not found: ID does not exist" Nov 22 08:31:33 crc kubenswrapper[4735]: I1122 08:31:33.278944 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" path="/var/lib/kubelet/pods/ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5/volumes" Nov 22 08:31:35 crc kubenswrapper[4735]: I1122 08:31:35.263236 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:31:35 crc kubenswrapper[4735]: E1122 08:31:35.263979 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:31:37 crc kubenswrapper[4735]: I1122 08:31:37.551265 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82tw9" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="registry-server" probeResult="failure" output=< Nov 22 08:31:37 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:31:37 crc kubenswrapper[4735]: > Nov 22 08:31:47 crc kubenswrapper[4735]: I1122 08:31:47.546901 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82tw9" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="registry-server" probeResult="failure" output=< Nov 22 08:31:47 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:31:47 crc kubenswrapper[4735]: > Nov 22 08:31:48 crc kubenswrapper[4735]: I1122 08:31:48.263499 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:31:48 crc kubenswrapper[4735]: E1122 08:31:48.263790 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:31:56 crc kubenswrapper[4735]: I1122 08:31:56.537839 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:56 crc kubenswrapper[4735]: I1122 08:31:56.605171 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:56 crc kubenswrapper[4735]: I1122 08:31:56.778742 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-82tw9"] Nov 22 08:31:57 crc kubenswrapper[4735]: I1122 08:31:57.680028 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-82tw9" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="registry-server" containerID="cri-o://ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c" gracePeriod=2 Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.340572 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.404574 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfd26\" (UniqueName: \"kubernetes.io/projected/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-kube-api-access-kfd26\") pod \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.404683 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-catalog-content\") pod \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.404851 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-utilities\") pod \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\" (UID: \"76304aea-23c2-43ca-a7f9-4a524f4c1bb3\") " Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.406125 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-utilities" (OuterVolumeSpecName: "utilities") pod "76304aea-23c2-43ca-a7f9-4a524f4c1bb3" (UID: "76304aea-23c2-43ca-a7f9-4a524f4c1bb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.416850 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-kube-api-access-kfd26" (OuterVolumeSpecName: "kube-api-access-kfd26") pod "76304aea-23c2-43ca-a7f9-4a524f4c1bb3" (UID: "76304aea-23c2-43ca-a7f9-4a524f4c1bb3"). InnerVolumeSpecName "kube-api-access-kfd26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.507139 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfd26\" (UniqueName: \"kubernetes.io/projected/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-kube-api-access-kfd26\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.507181 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.520776 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76304aea-23c2-43ca-a7f9-4a524f4c1bb3" (UID: "76304aea-23c2-43ca-a7f9-4a524f4c1bb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.611235 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76304aea-23c2-43ca-a7f9-4a524f4c1bb3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.698706 4735 generic.go:334] "Generic (PLEG): container finished" podID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerID="ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c" exitCode=0 Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.698804 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82tw9" event={"ID":"76304aea-23c2-43ca-a7f9-4a524f4c1bb3","Type":"ContainerDied","Data":"ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c"} Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.699009 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82tw9" event={"ID":"76304aea-23c2-43ca-a7f9-4a524f4c1bb3","Type":"ContainerDied","Data":"dc5dd491a2fd068059c7df3de6ad3cad5eb8baac9b7196eb36ed35d07297f9a5"} Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.699048 4735 scope.go:117] "RemoveContainer" containerID="ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.699500 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82tw9" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.760060 4735 scope.go:117] "RemoveContainer" containerID="49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.777158 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-82tw9"] Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.793523 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-82tw9"] Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.805604 4735 scope.go:117] "RemoveContainer" containerID="7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.857262 4735 scope.go:117] "RemoveContainer" containerID="ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c" Nov 22 08:31:58 crc kubenswrapper[4735]: E1122 08:31:58.858435 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c\": container with ID starting with ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c not found: ID does not exist" containerID="ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.858527 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c"} err="failed to get container status \"ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c\": rpc error: code = NotFound desc = could not find container \"ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c\": container with ID starting with ad7c36ea5d2cc172ab86d7599ce45560713a36d4739a651638632c9368878a1c not found: ID does not exist" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.858569 4735 scope.go:117] "RemoveContainer" containerID="49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf" Nov 22 08:31:58 crc kubenswrapper[4735]: E1122 08:31:58.859062 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf\": container with ID starting with 49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf not found: ID does not exist" containerID="49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.859107 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf"} err="failed to get container status \"49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf\": rpc error: code = NotFound desc = could not find container \"49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf\": container with ID starting with 49ee51d27ca7ca90292dea6319c59fef866b7f405352315426220ae3bec35edf not found: ID does not exist" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.859136 4735 scope.go:117] "RemoveContainer" containerID="7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b" Nov 22 08:31:58 crc kubenswrapper[4735]: E1122 08:31:58.859447 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b\": container with ID starting with 7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b not found: ID does not exist" containerID="7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b" Nov 22 08:31:58 crc kubenswrapper[4735]: I1122 08:31:58.859510 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b"} err="failed to get container status \"7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b\": rpc error: code = NotFound desc = could not find container \"7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b\": container with ID starting with 7879d55217d2bcd1e59e2a189f3e0d786a703afefda289f6829be0a6bd330b0b not found: ID does not exist" Nov 22 08:31:59 crc kubenswrapper[4735]: I1122 08:31:59.288568 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" path="/var/lib/kubelet/pods/76304aea-23c2-43ca-a7f9-4a524f4c1bb3/volumes" Nov 22 08:32:01 crc kubenswrapper[4735]: I1122 08:32:01.263828 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:32:01 crc kubenswrapper[4735]: E1122 08:32:01.264407 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.036641 4735 scope.go:117] "RemoveContainer" containerID="c149bb79a9740082c76c781ede88172aff86af44fbfbd02d97318e12f40634b1" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.067454 4735 scope.go:117] "RemoveContainer" containerID="c8ea1500e428226bdd507961f97a1f2fc9ed4a47730c05233c5f59f53d35e882" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.096016 4735 scope.go:117] "RemoveContainer" containerID="12e61705ef784d5c0396940a6a3476abc45478d982f44322ba112446fc102109" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.140742 4735 scope.go:117] "RemoveContainer" containerID="b469a9e6ff5a08892757f334bfbcbf8482dac49298107cbb11f5735e82507ae5" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.168946 4735 scope.go:117] "RemoveContainer" containerID="8ebb0c9292dcb1041f95e192c1b40e3b7a324dd2bb34acddf3dd4a05bf1feb2b" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.189553 4735 scope.go:117] "RemoveContainer" containerID="da36175fbcaf8cd07a106ded62d8e880079a7dc7476028ddda46700dba7b76e4" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.219225 4735 scope.go:117] "RemoveContainer" containerID="14b4b3e3dd41f2b911525d3192e343b6a2a67001193bb5212da8909f19a034cd" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.253633 4735 scope.go:117] "RemoveContainer" containerID="cbcb10f4e1ead4c3a62dcdb5ecc619d242a4847559407a08570669f7b6c82553" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.289132 4735 scope.go:117] "RemoveContainer" containerID="78dd47744ff42aee374d5ad49ae8eb2c62c176f49f2b64186fa01bebbeccd25b" Nov 22 08:32:05 crc kubenswrapper[4735]: I1122 08:32:05.313555 4735 scope.go:117] "RemoveContainer" containerID="76f50ffbaf53f10a808c31db0ca8c755187aee83b8aff0683d0c3c4a9382f667" Nov 22 08:32:14 crc kubenswrapper[4735]: I1122 08:32:14.265154 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:32:14 crc kubenswrapper[4735]: E1122 08:32:14.266522 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:32:28 crc kubenswrapper[4735]: I1122 08:32:28.263606 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:32:28 crc kubenswrapper[4735]: E1122 08:32:28.264441 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:32:40 crc kubenswrapper[4735]: I1122 08:32:40.263246 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:32:40 crc kubenswrapper[4735]: E1122 08:32:40.264002 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:32:53 crc kubenswrapper[4735]: I1122 08:32:53.264009 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:32:53 crc kubenswrapper[4735]: E1122 08:32:53.264791 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:33:06 crc kubenswrapper[4735]: I1122 08:33:06.263731 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:33:06 crc kubenswrapper[4735]: E1122 08:33:06.264536 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:33:19 crc kubenswrapper[4735]: I1122 08:33:19.284322 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:33:19 crc kubenswrapper[4735]: I1122 08:33:19.728587 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"4681e22fbcdfadcc11184336c0fc4ee081ea57446a881d6cbac7c3aa51dc6f7c"} Nov 22 08:33:21 crc kubenswrapper[4735]: I1122 08:33:21.759691 4735 generic.go:334] "Generic (PLEG): container finished" podID="d5c15b81-20c1-46c8-8f80-dcb4e7394735" containerID="29ddfc01175ae0a199f348892a4cc702e75fc8ea90962960649ac664fa627fc4" exitCode=0 Nov 22 08:33:21 crc kubenswrapper[4735]: I1122 08:33:21.759772 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" event={"ID":"d5c15b81-20c1-46c8-8f80-dcb4e7394735","Type":"ContainerDied","Data":"29ddfc01175ae0a199f348892a4cc702e75fc8ea90962960649ac664fa627fc4"} Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.266378 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.407041 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-bootstrap-combined-ca-bundle\") pod \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.407124 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq48b\" (UniqueName: \"kubernetes.io/projected/d5c15b81-20c1-46c8-8f80-dcb4e7394735-kube-api-access-sq48b\") pod \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.407258 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-inventory\") pod \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.407360 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-ssh-key\") pod \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\" (UID: \"d5c15b81-20c1-46c8-8f80-dcb4e7394735\") " Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.415828 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c15b81-20c1-46c8-8f80-dcb4e7394735-kube-api-access-sq48b" (OuterVolumeSpecName: "kube-api-access-sq48b") pod "d5c15b81-20c1-46c8-8f80-dcb4e7394735" (UID: "d5c15b81-20c1-46c8-8f80-dcb4e7394735"). InnerVolumeSpecName "kube-api-access-sq48b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.416481 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d5c15b81-20c1-46c8-8f80-dcb4e7394735" (UID: "d5c15b81-20c1-46c8-8f80-dcb4e7394735"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.450982 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d5c15b81-20c1-46c8-8f80-dcb4e7394735" (UID: "d5c15b81-20c1-46c8-8f80-dcb4e7394735"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.475524 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-inventory" (OuterVolumeSpecName: "inventory") pod "d5c15b81-20c1-46c8-8f80-dcb4e7394735" (UID: "d5c15b81-20c1-46c8-8f80-dcb4e7394735"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.510242 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.510290 4735 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.510311 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq48b\" (UniqueName: \"kubernetes.io/projected/d5c15b81-20c1-46c8-8f80-dcb4e7394735-kube-api-access-sq48b\") on node \"crc\" DevicePath \"\"" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.510329 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c15b81-20c1-46c8-8f80-dcb4e7394735-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.784738 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" event={"ID":"d5c15b81-20c1-46c8-8f80-dcb4e7394735","Type":"ContainerDied","Data":"5fe47b693ba640e509b6aa9fb0daf2ac046fad5d5de7f305b95a802ac4737a64"} Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.784788 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fe47b693ba640e509b6aa9fb0daf2ac046fad5d5de7f305b95a802ac4737a64" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.784842 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.909529 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw"] Nov 22 08:33:23 crc kubenswrapper[4735]: E1122 08:33:23.910410 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="extract-content" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910437 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="extract-content" Nov 22 08:33:23 crc kubenswrapper[4735]: E1122 08:33:23.910483 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="extract-utilities" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910517 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="extract-utilities" Nov 22 08:33:23 crc kubenswrapper[4735]: E1122 08:33:23.910541 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="registry-server" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910550 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="registry-server" Nov 22 08:33:23 crc kubenswrapper[4735]: E1122 08:33:23.910569 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="registry-server" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910578 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="registry-server" Nov 22 08:33:23 crc kubenswrapper[4735]: E1122 08:33:23.910603 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="extract-utilities" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910611 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="extract-utilities" Nov 22 08:33:23 crc kubenswrapper[4735]: E1122 08:33:23.910622 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="extract-content" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910630 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="extract-content" Nov 22 08:33:23 crc kubenswrapper[4735]: E1122 08:33:23.910649 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c15b81-20c1-46c8-8f80-dcb4e7394735" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910658 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c15b81-20c1-46c8-8f80-dcb4e7394735" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910957 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddee29d1-53a1-47f5-9a8b-89c2b3f60bb5" containerName="registry-server" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.910977 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c15b81-20c1-46c8-8f80-dcb4e7394735" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.911002 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="76304aea-23c2-43ca-a7f9-4a524f4c1bb3" containerName="registry-server" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.912040 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.914901 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.917240 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.917274 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.918415 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:33:23 crc kubenswrapper[4735]: I1122 08:33:23.944446 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.022823 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.022998 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9kpz\" (UniqueName: \"kubernetes.io/projected/8b249f3f-b08f-44b7-8652-9db500ef58b5-kube-api-access-k9kpz\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.023253 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.064275 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-2f1c-account-create-f4v9s"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.074868 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-2f1c-account-create-f4v9s"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.085965 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-nfpfq"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.097385 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c517-account-create-gg4qg"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.107272 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c517-account-create-gg4qg"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.116661 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-nfpfq"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.124599 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9kpz\" (UniqueName: \"kubernetes.io/projected/8b249f3f-b08f-44b7-8652-9db500ef58b5-kube-api-access-k9kpz\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.124705 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.124845 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.126830 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ztsfq"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.130965 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.131999 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.137793 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ztsfq"] Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.158067 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9kpz\" (UniqueName: \"kubernetes.io/projected/8b249f3f-b08f-44b7-8652-9db500ef58b5-kube-api-access-k9kpz\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.239012 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:33:24 crc kubenswrapper[4735]: I1122 08:33:24.806146 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw"] Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.030748 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2n7mj"] Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.042491 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f25a-account-create-rg9dx"] Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.053871 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2n7mj"] Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.074662 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f25a-account-create-rg9dx"] Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.308320 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e7957d0-44ed-4753-a100-253797cb1be9" path="/var/lib/kubelet/pods/4e7957d0-44ed-4753-a100-253797cb1be9/volumes" Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.309995 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="659efe09-eb81-4668-a68f-fc72b7bc43d1" path="/var/lib/kubelet/pods/659efe09-eb81-4668-a68f-fc72b7bc43d1/volumes" Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.317610 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e8014b5-baf9-4531-a275-2df3ba0c2af1" path="/var/lib/kubelet/pods/6e8014b5-baf9-4531-a275-2df3ba0c2af1/volumes" Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.319612 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77979ac0-cbb3-4785-9bd5-14cfb29749f3" path="/var/lib/kubelet/pods/77979ac0-cbb3-4785-9bd5-14cfb29749f3/volumes" Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.320270 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98303611-37f9-4fee-89fb-c4b58d51ff2f" path="/var/lib/kubelet/pods/98303611-37f9-4fee-89fb-c4b58d51ff2f/volumes" Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.332629 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf0ea764-b639-4959-8a27-e2f674be3329" path="/var/lib/kubelet/pods/cf0ea764-b639-4959-8a27-e2f674be3329/volumes" Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.810739 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" event={"ID":"8b249f3f-b08f-44b7-8652-9db500ef58b5","Type":"ContainerStarted","Data":"d55a35d3601c1583d940d0677ebb249cd4757931e8e78269128c5ccafad71f29"} Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.811950 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" event={"ID":"8b249f3f-b08f-44b7-8652-9db500ef58b5","Type":"ContainerStarted","Data":"22f298e263ee0c4c09451eb5bc63ef479b798b0ec4288858fb2b25154f93d314"} Nov 22 08:33:25 crc kubenswrapper[4735]: I1122 08:33:25.839352 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" podStartSLOduration=2.334986436 podStartE2EDuration="2.839334028s" podCreationTimestamp="2025-11-22 08:33:23 +0000 UTC" firstStartedPulling="2025-11-22 08:33:24.808764255 +0000 UTC m=+1826.413102860" lastFinishedPulling="2025-11-22 08:33:25.313111847 +0000 UTC m=+1826.917450452" observedRunningTime="2025-11-22 08:33:25.829355564 +0000 UTC m=+1827.433694179" watchObservedRunningTime="2025-11-22 08:33:25.839334028 +0000 UTC m=+1827.443672633" Nov 22 08:33:27 crc kubenswrapper[4735]: I1122 08:33:27.038174 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-20d8-account-create-crjxr"] Nov 22 08:33:27 crc kubenswrapper[4735]: I1122 08:33:27.055402 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-dmgvb"] Nov 22 08:33:27 crc kubenswrapper[4735]: I1122 08:33:27.068311 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-dmgvb"] Nov 22 08:33:27 crc kubenswrapper[4735]: I1122 08:33:27.081360 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-20d8-account-create-crjxr"] Nov 22 08:33:27 crc kubenswrapper[4735]: I1122 08:33:27.284511 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f4ab0e4-1b2f-4c1a-8414-d85c38509a64" path="/var/lib/kubelet/pods/3f4ab0e4-1b2f-4c1a-8414-d85c38509a64/volumes" Nov 22 08:33:27 crc kubenswrapper[4735]: I1122 08:33:27.286035 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5e4fc11-5ae0-4336-bc26-f07ca5230b48" path="/var/lib/kubelet/pods/c5e4fc11-5ae0-4336-bc26-f07ca5230b48/volumes" Nov 22 08:33:32 crc kubenswrapper[4735]: I1122 08:33:32.043251 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-85db-account-create-rt6xj"] Nov 22 08:33:32 crc kubenswrapper[4735]: I1122 08:33:32.056598 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-8drgk"] Nov 22 08:33:32 crc kubenswrapper[4735]: I1122 08:33:32.070508 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-85db-account-create-rt6xj"] Nov 22 08:33:32 crc kubenswrapper[4735]: I1122 08:33:32.084180 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-8drgk"] Nov 22 08:33:33 crc kubenswrapper[4735]: I1122 08:33:33.283100 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a31925-48f6-4fba-91e0-b5884bde3344" path="/var/lib/kubelet/pods/30a31925-48f6-4fba-91e0-b5884bde3344/volumes" Nov 22 08:33:33 crc kubenswrapper[4735]: I1122 08:33:33.285588 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6d00bef-4c5e-40f0-9c2f-8811da61708c" path="/var/lib/kubelet/pods/c6d00bef-4c5e-40f0-9c2f-8811da61708c/volumes" Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.055442 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-098a-account-create-7sqdp"] Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.067578 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-9cd9-account-create-zxr8b"] Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.099258 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-098a-account-create-7sqdp"] Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.110055 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-bkgcz"] Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.119100 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-7cghr"] Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.131947 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-9cd9-account-create-zxr8b"] Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.142384 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-7cghr"] Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.151568 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-bkgcz"] Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.282758 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45fc366f-ea6e-4134-afa2-2612f5f0d5f3" path="/var/lib/kubelet/pods/45fc366f-ea6e-4134-afa2-2612f5f0d5f3/volumes" Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.284873 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4925fb22-c2bb-4609-a593-5589e9137ecc" path="/var/lib/kubelet/pods/4925fb22-c2bb-4609-a593-5589e9137ecc/volumes" Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.285892 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69db05e1-3ac1-4aaf-a558-6e8581dfbf2e" path="/var/lib/kubelet/pods/69db05e1-3ac1-4aaf-a558-6e8581dfbf2e/volumes" Nov 22 08:33:51 crc kubenswrapper[4735]: I1122 08:33:51.286833 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebfceb98-899b-40d4-a27d-dfb27042557e" path="/var/lib/kubelet/pods/ebfceb98-899b-40d4-a27d-dfb27042557e/volumes" Nov 22 08:33:54 crc kubenswrapper[4735]: I1122 08:33:54.058214 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3454-account-create-j5ssm"] Nov 22 08:33:54 crc kubenswrapper[4735]: I1122 08:33:54.075853 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-t9t64"] Nov 22 08:33:54 crc kubenswrapper[4735]: I1122 08:33:54.089912 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-t9t64"] Nov 22 08:33:54 crc kubenswrapper[4735]: I1122 08:33:54.099585 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3454-account-create-j5ssm"] Nov 22 08:33:54 crc kubenswrapper[4735]: I1122 08:33:54.109017 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-134f-account-create-g9cqb"] Nov 22 08:33:54 crc kubenswrapper[4735]: I1122 08:33:54.117611 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-8brw8"] Nov 22 08:33:54 crc kubenswrapper[4735]: I1122 08:33:54.126108 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-8brw8"] Nov 22 08:33:54 crc kubenswrapper[4735]: I1122 08:33:54.135374 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-134f-account-create-g9cqb"] Nov 22 08:33:55 crc kubenswrapper[4735]: I1122 08:33:55.279779 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="422ffe22-0786-4688-bd88-9e28481c59ff" path="/var/lib/kubelet/pods/422ffe22-0786-4688-bd88-9e28481c59ff/volumes" Nov 22 08:33:55 crc kubenswrapper[4735]: I1122 08:33:55.280850 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a566e05-aacd-44f9-b3dd-231549b9a967" path="/var/lib/kubelet/pods/5a566e05-aacd-44f9-b3dd-231549b9a967/volumes" Nov 22 08:33:55 crc kubenswrapper[4735]: I1122 08:33:55.283081 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7" path="/var/lib/kubelet/pods/8f4f41b7-19f8-4520-9d4f-9e1c67aeeae7/volumes" Nov 22 08:33:55 crc kubenswrapper[4735]: I1122 08:33:55.284026 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6331445-03fa-4581-a892-974fdd4e53fa" path="/var/lib/kubelet/pods/c6331445-03fa-4581-a892-974fdd4e53fa/volumes" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.481115 4735 scope.go:117] "RemoveContainer" containerID="9d4cc87de46a3e403e4260565b03edabe0f0fe1931783dfc01b982dbd84a32bb" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.508400 4735 scope.go:117] "RemoveContainer" containerID="e787fd4b999aade029525ca31150f0ca69d81e74d922b10588eea746b41a8001" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.622129 4735 scope.go:117] "RemoveContainer" containerID="0b072957e6f7373a5320c3e2a92f347bb471db0944c469c4e92d077ccc93a5e0" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.661093 4735 scope.go:117] "RemoveContainer" containerID="6ed26bbccf97073cf14b2cfbbe6e9da1892c4f89b3bc9b736a52aaa8d5d986f5" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.707992 4735 scope.go:117] "RemoveContainer" containerID="fc3c3652534f48f5bfe73cefe808e25bf3d0c43f4d74ffdb893b753719480640" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.763094 4735 scope.go:117] "RemoveContainer" containerID="51567db7aa746db0f673c4bd05a1613a8503972d79f08a7a09e0acfc13a81172" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.816858 4735 scope.go:117] "RemoveContainer" containerID="191e5ae3b96eca02d4554ac81a431b766af329bde7f9491675c3d30532a6b393" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.840006 4735 scope.go:117] "RemoveContainer" containerID="f85ac7ec786dbd469571c9691a06c4c6bc7cda82287dd0591119f0c7d1c3510c" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.869415 4735 scope.go:117] "RemoveContainer" containerID="8708cea40636f856f6bf29e50ac1ba1056fb28cd103ea3cf009c4a29eaebb3d9" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.899014 4735 scope.go:117] "RemoveContainer" containerID="1edd19d6822f1ac35e33e83fc3835af064231a438259df80b686b83e4d16e74c" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.921054 4735 scope.go:117] "RemoveContainer" containerID="aeef6bab3bffb95e65aa4f8a85045f2ddbe112332a288077632042e0ab193708" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.952308 4735 scope.go:117] "RemoveContainer" containerID="406a5f6bf3dcb697458cd3bd8db577b52baa9653d96a6c8c308d44f0a457e919" Nov 22 08:34:05 crc kubenswrapper[4735]: I1122 08:34:05.984142 4735 scope.go:117] "RemoveContainer" containerID="69c97d42e592285727803bccf2ff3a143a24f89af4233304a34858d0578bf42a" Nov 22 08:34:06 crc kubenswrapper[4735]: I1122 08:34:06.019866 4735 scope.go:117] "RemoveContainer" containerID="df8ad461779002e80910d31d5dffe0fcbe435f4f1cde76d276ed7f11d6ae4d10" Nov 22 08:34:06 crc kubenswrapper[4735]: I1122 08:34:06.054535 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-9grv6"] Nov 22 08:34:06 crc kubenswrapper[4735]: I1122 08:34:06.060168 4735 scope.go:117] "RemoveContainer" containerID="07bc0ff0794476d96188f8305cd08785cb95998803e5ce2165a534bb8edd86b7" Nov 22 08:34:06 crc kubenswrapper[4735]: I1122 08:34:06.070797 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-9grv6"] Nov 22 08:34:06 crc kubenswrapper[4735]: I1122 08:34:06.083713 4735 scope.go:117] "RemoveContainer" containerID="4856d1509ff777af929d217195a1f000a5c6a0ea2ac36c72a57d4e47b30fd734" Nov 22 08:34:06 crc kubenswrapper[4735]: I1122 08:34:06.106116 4735 scope.go:117] "RemoveContainer" containerID="3f3c694cb46e4f85d2d5c053df886a6d1f1e4b54dd01578dbecbe67748a0d446" Nov 22 08:34:06 crc kubenswrapper[4735]: I1122 08:34:06.133752 4735 scope.go:117] "RemoveContainer" containerID="27d9ba4ee2fcd61435f2e708644df7ef6daba680a65e417058ef81995c759113" Nov 22 08:34:06 crc kubenswrapper[4735]: I1122 08:34:06.159435 4735 scope.go:117] "RemoveContainer" containerID="a5bcee5f440f8e179c2ad85fc45661bf74347ae9960f818494e1ba614685722d" Nov 22 08:34:07 crc kubenswrapper[4735]: I1122 08:34:07.277982 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d75e4f-e760-41d2-b7d8-b9baffb60bcf" path="/var/lib/kubelet/pods/36d75e4f-e760-41d2-b7d8-b9baffb60bcf/volumes" Nov 22 08:34:36 crc kubenswrapper[4735]: I1122 08:34:36.054579 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-m9x5r"] Nov 22 08:34:36 crc kubenswrapper[4735]: I1122 08:34:36.068713 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-m9x5r"] Nov 22 08:34:37 crc kubenswrapper[4735]: I1122 08:34:37.282577 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54265b3d-7ff9-4b71-aab0-9b79ae0b08d3" path="/var/lib/kubelet/pods/54265b3d-7ff9-4b71-aab0-9b79ae0b08d3/volumes" Nov 22 08:34:47 crc kubenswrapper[4735]: I1122 08:34:47.037683 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-qptkr"] Nov 22 08:34:47 crc kubenswrapper[4735]: I1122 08:34:47.045616 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-qptkr"] Nov 22 08:34:47 crc kubenswrapper[4735]: I1122 08:34:47.288689 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f525dad0-8aeb-4928-b87f-561f2c3f3228" path="/var/lib/kubelet/pods/f525dad0-8aeb-4928-b87f-561f2c3f3228/volumes" Nov 22 08:34:48 crc kubenswrapper[4735]: I1122 08:34:48.033403 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-t9c5x"] Nov 22 08:34:48 crc kubenswrapper[4735]: I1122 08:34:48.045285 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nj6dc"] Nov 22 08:34:48 crc kubenswrapper[4735]: I1122 08:34:48.055267 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-t9c5x"] Nov 22 08:34:48 crc kubenswrapper[4735]: I1122 08:34:48.065206 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nj6dc"] Nov 22 08:34:49 crc kubenswrapper[4735]: I1122 08:34:49.276624 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32d2023-853f-40f2-9322-155b32c82fdb" path="/var/lib/kubelet/pods/a32d2023-853f-40f2-9322-155b32c82fdb/volumes" Nov 22 08:34:49 crc kubenswrapper[4735]: I1122 08:34:49.277628 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a560c2a6-316e-4d1c-8b84-4a44aec7e772" path="/var/lib/kubelet/pods/a560c2a6-316e-4d1c-8b84-4a44aec7e772/volumes" Nov 22 08:34:50 crc kubenswrapper[4735]: I1122 08:34:50.043953 4735 patch_prober.go:28] interesting pod/logging-loki-gateway-6f9c8b877-5dtcw container/gateway namespace/openshift-logging: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={ Nov 22 08:34:50 crc kubenswrapper[4735]: "http": "returned status 503, expected 200" Nov 22 08:34:50 crc kubenswrapper[4735]: } Nov 22 08:34:50 crc kubenswrapper[4735]: I1122 08:34:50.044260 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" podUID="5caa9595-315a-40a7-a46f-5e6d4b13c320" containerName="gateway" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 22 08:34:56 crc kubenswrapper[4735]: I1122 08:34:56.019348 4735 patch_prober.go:28] interesting pod/logging-loki-gateway-6f9c8b877-jv5cr container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.76:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 08:34:56 crc kubenswrapper[4735]: I1122 08:34:56.020102 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" podUID="3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.76:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 08:35:01 crc kubenswrapper[4735]: I1122 08:35:01.049265 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-z64cd"] Nov 22 08:35:01 crc kubenswrapper[4735]: I1122 08:35:01.061714 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-z64cd"] Nov 22 08:35:01 crc kubenswrapper[4735]: I1122 08:35:01.276356 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e519a2d0-063e-4097-80f8-f4276208317a" path="/var/lib/kubelet/pods/e519a2d0-063e-4097-80f8-f4276208317a/volumes" Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.043741 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-qsp4r"] Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.058883 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-qsp4r"] Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.626840 4735 scope.go:117] "RemoveContainer" containerID="82367da6640e3280143e869e40240f2b5acad22fd6e4644929227988c5465988" Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.739732 4735 scope.go:117] "RemoveContainer" containerID="e8cbdd89932353d5e2975889e319e29e0af3f59f699ee85b488473c10d8fba74" Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.784911 4735 scope.go:117] "RemoveContainer" containerID="0ce4edff93f5c6b558eda4f8d79c20719809371092b22ec4240d53027338cab9" Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.831221 4735 scope.go:117] "RemoveContainer" containerID="6462b6967208bd3cca5d8300b6b254224c25e883fc9eaa70d61b234d15c25876" Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.874170 4735 scope.go:117] "RemoveContainer" containerID="2762bf4a14beb7350f245c7468efcceed103ef8107b01e8b3517f2926729d4f4" Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.894743 4735 scope.go:117] "RemoveContainer" containerID="ae00901a5c9b52914ee48e511f5380bb1657eb56dee8e3af32ea8fce46538b32" Nov 22 08:35:06 crc kubenswrapper[4735]: I1122 08:35:06.964397 4735 scope.go:117] "RemoveContainer" containerID="427d0f580305583837f5847a430716a462eec9578979f975200112fc0c9a452c" Nov 22 08:35:07 crc kubenswrapper[4735]: I1122 08:35:07.084879 4735 scope.go:117] "RemoveContainer" containerID="fa99cdc915a3e62de9d4cb9945e176041e12d5b45dc2555fe50a48563aeccab1" Nov 22 08:35:07 crc kubenswrapper[4735]: I1122 08:35:07.121299 4735 scope.go:117] "RemoveContainer" containerID="1564a2ea10c090dbdeccf918d54f3a7fa0082ac5ca522861ca861a839b46bf97" Nov 22 08:35:07 crc kubenswrapper[4735]: I1122 08:35:07.191951 4735 scope.go:117] "RemoveContainer" containerID="0798dc5eda46a091c75461a2bb9584d339a13a861f521d088d8295c91440a1d4" Nov 22 08:35:07 crc kubenswrapper[4735]: I1122 08:35:07.278343 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9f75ef9-02e1-4af0-b532-2e744d02d842" path="/var/lib/kubelet/pods/c9f75ef9-02e1-4af0-b532-2e744d02d842/volumes" Nov 22 08:35:24 crc kubenswrapper[4735]: I1122 08:35:24.983226 4735 patch_prober.go:28] interesting pod/logging-loki-gateway-6f9c8b877-jv5cr container/gateway namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={ Nov 22 08:35:24 crc kubenswrapper[4735]: "http": "returned status 503, expected 200" Nov 22 08:35:24 crc kubenswrapper[4735]: } Nov 22 08:35:24 crc kubenswrapper[4735]: I1122 08:35:24.983806 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" podUID="3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0" containerName="gateway" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 22 08:35:27 crc kubenswrapper[4735]: I1122 08:35:27.454991 4735 generic.go:334] "Generic (PLEG): container finished" podID="8b249f3f-b08f-44b7-8652-9db500ef58b5" containerID="d55a35d3601c1583d940d0677ebb249cd4757931e8e78269128c5ccafad71f29" exitCode=0 Nov 22 08:35:27 crc kubenswrapper[4735]: I1122 08:35:27.455093 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" event={"ID":"8b249f3f-b08f-44b7-8652-9db500ef58b5","Type":"ContainerDied","Data":"d55a35d3601c1583d940d0677ebb249cd4757931e8e78269128c5ccafad71f29"} Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.071594 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.193775 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9kpz\" (UniqueName: \"kubernetes.io/projected/8b249f3f-b08f-44b7-8652-9db500ef58b5-kube-api-access-k9kpz\") pod \"8b249f3f-b08f-44b7-8652-9db500ef58b5\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.193851 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-inventory\") pod \"8b249f3f-b08f-44b7-8652-9db500ef58b5\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.194067 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-ssh-key\") pod \"8b249f3f-b08f-44b7-8652-9db500ef58b5\" (UID: \"8b249f3f-b08f-44b7-8652-9db500ef58b5\") " Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.201266 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b249f3f-b08f-44b7-8652-9db500ef58b5-kube-api-access-k9kpz" (OuterVolumeSpecName: "kube-api-access-k9kpz") pod "8b249f3f-b08f-44b7-8652-9db500ef58b5" (UID: "8b249f3f-b08f-44b7-8652-9db500ef58b5"). InnerVolumeSpecName "kube-api-access-k9kpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.232049 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8b249f3f-b08f-44b7-8652-9db500ef58b5" (UID: "8b249f3f-b08f-44b7-8652-9db500ef58b5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.241830 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-inventory" (OuterVolumeSpecName: "inventory") pod "8b249f3f-b08f-44b7-8652-9db500ef58b5" (UID: "8b249f3f-b08f-44b7-8652-9db500ef58b5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.297223 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.297256 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9kpz\" (UniqueName: \"kubernetes.io/projected/8b249f3f-b08f-44b7-8652-9db500ef58b5-kube-api-access-k9kpz\") on node \"crc\" DevicePath \"\"" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.297268 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b249f3f-b08f-44b7-8652-9db500ef58b5-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.480417 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" event={"ID":"8b249f3f-b08f-44b7-8652-9db500ef58b5","Type":"ContainerDied","Data":"22f298e263ee0c4c09451eb5bc63ef479b798b0ec4288858fb2b25154f93d314"} Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.480476 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22f298e263ee0c4c09451eb5bc63ef479b798b0ec4288858fb2b25154f93d314" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.480510 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.562635 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7"] Nov 22 08:35:29 crc kubenswrapper[4735]: E1122 08:35:29.563730 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b249f3f-b08f-44b7-8652-9db500ef58b5" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.563761 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b249f3f-b08f-44b7-8652-9db500ef58b5" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.564139 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b249f3f-b08f-44b7-8652-9db500ef58b5" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.565322 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.567396 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.568190 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.568640 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.568895 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.592848 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7"] Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.731214 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq6p7\" (UniqueName: \"kubernetes.io/projected/ab4508bb-701a-4c44-bdea-5f3e176afe47-kube-api-access-hq6p7\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.731349 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.732027 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.833643 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.833782 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq6p7\" (UniqueName: \"kubernetes.io/projected/ab4508bb-701a-4c44-bdea-5f3e176afe47-kube-api-access-hq6p7\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.833840 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.840344 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.840771 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.853073 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq6p7\" (UniqueName: \"kubernetes.io/projected/ab4508bb-701a-4c44-bdea-5f3e176afe47-kube-api-access-hq6p7\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:29 crc kubenswrapper[4735]: I1122 08:35:29.890524 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:35:30 crc kubenswrapper[4735]: I1122 08:35:30.542134 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7"] Nov 22 08:35:30 crc kubenswrapper[4735]: I1122 08:35:30.548599 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:35:31 crc kubenswrapper[4735]: I1122 08:35:31.503528 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" event={"ID":"ab4508bb-701a-4c44-bdea-5f3e176afe47","Type":"ContainerStarted","Data":"71b9c5805c221cb132b795022753e208f74031df300fa6f35360403ad89dcc21"} Nov 22 08:35:32 crc kubenswrapper[4735]: I1122 08:35:32.526664 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" event={"ID":"ab4508bb-701a-4c44-bdea-5f3e176afe47","Type":"ContainerStarted","Data":"d41694e02b2dfddd1095a211ef534f521ce3a48359d9fd116392344dc4b5be33"} Nov 22 08:35:32 crc kubenswrapper[4735]: I1122 08:35:32.565344 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" podStartSLOduration=2.805529995 podStartE2EDuration="3.565318999s" podCreationTimestamp="2025-11-22 08:35:29 +0000 UTC" firstStartedPulling="2025-11-22 08:35:30.548312255 +0000 UTC m=+1952.152650860" lastFinishedPulling="2025-11-22 08:35:31.308101239 +0000 UTC m=+1952.912439864" observedRunningTime="2025-11-22 08:35:32.548508075 +0000 UTC m=+1954.152846700" watchObservedRunningTime="2025-11-22 08:35:32.565318999 +0000 UTC m=+1954.169657624" Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.074848 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-npss9"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.087237 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-16cf-account-create-7jcz2"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.101548 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-7129-account-create-8nxqz"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.113656 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-89zq6"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.128402 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pl47j"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.139260 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-d80d-account-create-lcmww"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.150688 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-16cf-account-create-7jcz2"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.162769 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-89zq6"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.173924 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-d80d-account-create-lcmww"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.185954 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pl47j"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.197626 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-npss9"] Nov 22 08:35:40 crc kubenswrapper[4735]: I1122 08:35:40.209718 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-7129-account-create-8nxqz"] Nov 22 08:35:41 crc kubenswrapper[4735]: I1122 08:35:41.279162 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27d5c57d-8359-4b2f-96e5-d42f40634fd1" path="/var/lib/kubelet/pods/27d5c57d-8359-4b2f-96e5-d42f40634fd1/volumes" Nov 22 08:35:41 crc kubenswrapper[4735]: I1122 08:35:41.281069 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88" path="/var/lib/kubelet/pods/345b5d97-4fcc-4ba2-b2d5-c5ea347a2c88/volumes" Nov 22 08:35:41 crc kubenswrapper[4735]: I1122 08:35:41.289250 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b854f12-c3ca-424a-8c7d-12d4d708174f" path="/var/lib/kubelet/pods/8b854f12-c3ca-424a-8c7d-12d4d708174f/volumes" Nov 22 08:35:41 crc kubenswrapper[4735]: I1122 08:35:41.290731 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="991c0a20-2e5f-441a-ac7f-a74d0a83814e" path="/var/lib/kubelet/pods/991c0a20-2e5f-441a-ac7f-a74d0a83814e/volumes" Nov 22 08:35:41 crc kubenswrapper[4735]: I1122 08:35:41.292125 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0eb1004-ded0-44d0-8341-f51eedd26af3" path="/var/lib/kubelet/pods/a0eb1004-ded0-44d0-8341-f51eedd26af3/volumes" Nov 22 08:35:41 crc kubenswrapper[4735]: I1122 08:35:41.294917 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7bec6aa-ba3f-495e-bd32-52eb888657ac" path="/var/lib/kubelet/pods/d7bec6aa-ba3f-495e-bd32-52eb888657ac/volumes" Nov 22 08:35:46 crc kubenswrapper[4735]: I1122 08:35:46.131398 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:35:46 crc kubenswrapper[4735]: I1122 08:35:46.131871 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:36:07 crc kubenswrapper[4735]: I1122 08:36:07.483806 4735 scope.go:117] "RemoveContainer" containerID="e11ce79e7c417e089cbd02e1985b969c78b556e5bf67d6b391ff280846122cf8" Nov 22 08:36:07 crc kubenswrapper[4735]: I1122 08:36:07.516662 4735 scope.go:117] "RemoveContainer" containerID="bd0b109ae928074b6c650543bfc8b22b8acdabc52080177050d1e17b716fadfb" Nov 22 08:36:07 crc kubenswrapper[4735]: I1122 08:36:07.625230 4735 scope.go:117] "RemoveContainer" containerID="29d6a4d2a95babc7f17622223035ec852d5c4ccd92902d75abc1b039c548c691" Nov 22 08:36:07 crc kubenswrapper[4735]: I1122 08:36:07.670978 4735 scope.go:117] "RemoveContainer" containerID="6a111f241484eb5391838b9af58a78e8bf22389fb19a1d2438956734070d506d" Nov 22 08:36:07 crc kubenswrapper[4735]: I1122 08:36:07.749765 4735 scope.go:117] "RemoveContainer" containerID="9904d55dc8b1dee2eafe426aa9211ddcb4efb7cc5fe879cab7931d1ecb0659ed" Nov 22 08:36:07 crc kubenswrapper[4735]: I1122 08:36:07.795451 4735 scope.go:117] "RemoveContainer" containerID="f94fa5d7222201202d8524076d62ab837ec52117dcc4bb104cd31b816c7731ca" Nov 22 08:36:07 crc kubenswrapper[4735]: I1122 08:36:07.850546 4735 scope.go:117] "RemoveContainer" containerID="fea186d92751736a4ba2f572e095a82e7d5c1ea21e90b189cf09b7ed91f41b8b" Nov 22 08:36:16 crc kubenswrapper[4735]: I1122 08:36:16.053058 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sjwjn"] Nov 22 08:36:16 crc kubenswrapper[4735]: I1122 08:36:16.066810 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sjwjn"] Nov 22 08:36:16 crc kubenswrapper[4735]: I1122 08:36:16.141439 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:36:16 crc kubenswrapper[4735]: I1122 08:36:16.142012 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:36:17 crc kubenswrapper[4735]: I1122 08:36:17.278961 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c3a99da-447d-4586-b5d8-2214d6df2cbd" path="/var/lib/kubelet/pods/6c3a99da-447d-4586-b5d8-2214d6df2cbd/volumes" Nov 22 08:36:44 crc kubenswrapper[4735]: I1122 08:36:44.053738 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-nl8pr"] Nov 22 08:36:44 crc kubenswrapper[4735]: I1122 08:36:44.065271 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-nl8pr"] Nov 22 08:36:45 crc kubenswrapper[4735]: I1122 08:36:45.033111 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5brff"] Nov 22 08:36:45 crc kubenswrapper[4735]: I1122 08:36:45.048116 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5brff"] Nov 22 08:36:45 crc kubenswrapper[4735]: I1122 08:36:45.289030 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="581f275c-c454-46b5-8775-c10a814cfd4c" path="/var/lib/kubelet/pods/581f275c-c454-46b5-8775-c10a814cfd4c/volumes" Nov 22 08:36:45 crc kubenswrapper[4735]: I1122 08:36:45.321900 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d5067c-9432-415b-a20b-d0822ae72701" path="/var/lib/kubelet/pods/b3d5067c-9432-415b-a20b-d0822ae72701/volumes" Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.132181 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.132539 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.132599 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.133526 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4681e22fbcdfadcc11184336c0fc4ee081ea57446a881d6cbac7c3aa51dc6f7c"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.133601 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://4681e22fbcdfadcc11184336c0fc4ee081ea57446a881d6cbac7c3aa51dc6f7c" gracePeriod=600 Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.497216 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="4681e22fbcdfadcc11184336c0fc4ee081ea57446a881d6cbac7c3aa51dc6f7c" exitCode=0 Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.497317 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"4681e22fbcdfadcc11184336c0fc4ee081ea57446a881d6cbac7c3aa51dc6f7c"} Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.497386 4735 scope.go:117] "RemoveContainer" containerID="2e52ace2ea94f98f6c03ad7d89aa658a2221c95183660fe4358239c497d594f7" Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.499287 4735 generic.go:334] "Generic (PLEG): container finished" podID="ab4508bb-701a-4c44-bdea-5f3e176afe47" containerID="d41694e02b2dfddd1095a211ef534f521ce3a48359d9fd116392344dc4b5be33" exitCode=0 Nov 22 08:36:46 crc kubenswrapper[4735]: I1122 08:36:46.499322 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" event={"ID":"ab4508bb-701a-4c44-bdea-5f3e176afe47","Type":"ContainerDied","Data":"d41694e02b2dfddd1095a211ef534f521ce3a48359d9fd116392344dc4b5be33"} Nov 22 08:36:47 crc kubenswrapper[4735]: I1122 08:36:47.514555 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9"} Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.077057 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.214868 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-ssh-key\") pod \"ab4508bb-701a-4c44-bdea-5f3e176afe47\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.215090 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq6p7\" (UniqueName: \"kubernetes.io/projected/ab4508bb-701a-4c44-bdea-5f3e176afe47-kube-api-access-hq6p7\") pod \"ab4508bb-701a-4c44-bdea-5f3e176afe47\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.215282 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-inventory\") pod \"ab4508bb-701a-4c44-bdea-5f3e176afe47\" (UID: \"ab4508bb-701a-4c44-bdea-5f3e176afe47\") " Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.225156 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4508bb-701a-4c44-bdea-5f3e176afe47-kube-api-access-hq6p7" (OuterVolumeSpecName: "kube-api-access-hq6p7") pod "ab4508bb-701a-4c44-bdea-5f3e176afe47" (UID: "ab4508bb-701a-4c44-bdea-5f3e176afe47"). InnerVolumeSpecName "kube-api-access-hq6p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.245969 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-inventory" (OuterVolumeSpecName: "inventory") pod "ab4508bb-701a-4c44-bdea-5f3e176afe47" (UID: "ab4508bb-701a-4c44-bdea-5f3e176afe47"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.268087 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ab4508bb-701a-4c44-bdea-5f3e176afe47" (UID: "ab4508bb-701a-4c44-bdea-5f3e176afe47"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.318261 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq6p7\" (UniqueName: \"kubernetes.io/projected/ab4508bb-701a-4c44-bdea-5f3e176afe47-kube-api-access-hq6p7\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.318310 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.318322 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4508bb-701a-4c44-bdea-5f3e176afe47-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.527613 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.527616 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7" event={"ID":"ab4508bb-701a-4c44-bdea-5f3e176afe47","Type":"ContainerDied","Data":"71b9c5805c221cb132b795022753e208f74031df300fa6f35360403ad89dcc21"} Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.527872 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71b9c5805c221cb132b795022753e208f74031df300fa6f35360403ad89dcc21" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.645228 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf"] Nov 22 08:36:48 crc kubenswrapper[4735]: E1122 08:36:48.646060 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4508bb-701a-4c44-bdea-5f3e176afe47" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.646094 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4508bb-701a-4c44-bdea-5f3e176afe47" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.646515 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4508bb-701a-4c44-bdea-5f3e176afe47" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.647878 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.650377 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.650389 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.650756 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.656713 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.660892 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf"] Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.726205 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.726300 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.726356 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27xx4\" (UniqueName: \"kubernetes.io/projected/00a5290b-dff8-438f-a047-33f1d5f05f4d-kube-api-access-27xx4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.829036 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.829131 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.829184 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27xx4\" (UniqueName: \"kubernetes.io/projected/00a5290b-dff8-438f-a047-33f1d5f05f4d-kube-api-access-27xx4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.838082 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.846934 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.878483 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27xx4\" (UniqueName: \"kubernetes.io/projected/00a5290b-dff8-438f-a047-33f1d5f05f4d-kube-api-access-27xx4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-76smf\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:48 crc kubenswrapper[4735]: I1122 08:36:48.971855 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:49 crc kubenswrapper[4735]: I1122 08:36:49.645696 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf"] Nov 22 08:36:49 crc kubenswrapper[4735]: W1122 08:36:49.682943 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a5290b_dff8_438f_a047_33f1d5f05f4d.slice/crio-11bc5859f1ffc00c8eff2f17498a0daa04b3fa60379e483d0a9c353e0fa90c70 WatchSource:0}: Error finding container 11bc5859f1ffc00c8eff2f17498a0daa04b3fa60379e483d0a9c353e0fa90c70: Status 404 returned error can't find the container with id 11bc5859f1ffc00c8eff2f17498a0daa04b3fa60379e483d0a9c353e0fa90c70 Nov 22 08:36:50 crc kubenswrapper[4735]: I1122 08:36:50.549361 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" event={"ID":"00a5290b-dff8-438f-a047-33f1d5f05f4d","Type":"ContainerStarted","Data":"e36ff70118fd5d8c9a1603b576871800724f4d0a3384e0266cd1a42d549b7e90"} Nov 22 08:36:50 crc kubenswrapper[4735]: I1122 08:36:50.550020 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" event={"ID":"00a5290b-dff8-438f-a047-33f1d5f05f4d","Type":"ContainerStarted","Data":"11bc5859f1ffc00c8eff2f17498a0daa04b3fa60379e483d0a9c353e0fa90c70"} Nov 22 08:36:50 crc kubenswrapper[4735]: I1122 08:36:50.568135 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" podStartSLOduration=2.13141718 podStartE2EDuration="2.56811274s" podCreationTimestamp="2025-11-22 08:36:48 +0000 UTC" firstStartedPulling="2025-11-22 08:36:49.685215309 +0000 UTC m=+2031.289553914" lastFinishedPulling="2025-11-22 08:36:50.121910849 +0000 UTC m=+2031.726249474" observedRunningTime="2025-11-22 08:36:50.566985212 +0000 UTC m=+2032.171323827" watchObservedRunningTime="2025-11-22 08:36:50.56811274 +0000 UTC m=+2032.172451345" Nov 22 08:36:55 crc kubenswrapper[4735]: I1122 08:36:55.615651 4735 generic.go:334] "Generic (PLEG): container finished" podID="00a5290b-dff8-438f-a047-33f1d5f05f4d" containerID="e36ff70118fd5d8c9a1603b576871800724f4d0a3384e0266cd1a42d549b7e90" exitCode=0 Nov 22 08:36:55 crc kubenswrapper[4735]: I1122 08:36:55.615706 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" event={"ID":"00a5290b-dff8-438f-a047-33f1d5f05f4d","Type":"ContainerDied","Data":"e36ff70118fd5d8c9a1603b576871800724f4d0a3384e0266cd1a42d549b7e90"} Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.049125 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-8ebf-account-create-6ptmj"] Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.060866 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-8ebf-account-create-6ptmj"] Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.191881 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.242824 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-ssh-key\") pod \"00a5290b-dff8-438f-a047-33f1d5f05f4d\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.242948 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-inventory\") pod \"00a5290b-dff8-438f-a047-33f1d5f05f4d\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.243027 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27xx4\" (UniqueName: \"kubernetes.io/projected/00a5290b-dff8-438f-a047-33f1d5f05f4d-kube-api-access-27xx4\") pod \"00a5290b-dff8-438f-a047-33f1d5f05f4d\" (UID: \"00a5290b-dff8-438f-a047-33f1d5f05f4d\") " Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.250025 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a5290b-dff8-438f-a047-33f1d5f05f4d-kube-api-access-27xx4" (OuterVolumeSpecName: "kube-api-access-27xx4") pod "00a5290b-dff8-438f-a047-33f1d5f05f4d" (UID: "00a5290b-dff8-438f-a047-33f1d5f05f4d"). InnerVolumeSpecName "kube-api-access-27xx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.276776 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "00a5290b-dff8-438f-a047-33f1d5f05f4d" (UID: "00a5290b-dff8-438f-a047-33f1d5f05f4d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.284134 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090" path="/var/lib/kubelet/pods/ed5c8b5c-bc85-4546-9ff1-bd1f0f6e5090/volumes" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.310377 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-inventory" (OuterVolumeSpecName: "inventory") pod "00a5290b-dff8-438f-a047-33f1d5f05f4d" (UID: "00a5290b-dff8-438f-a047-33f1d5f05f4d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.345566 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27xx4\" (UniqueName: \"kubernetes.io/projected/00a5290b-dff8-438f-a047-33f1d5f05f4d-kube-api-access-27xx4\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.345617 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.345635 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00a5290b-dff8-438f-a047-33f1d5f05f4d-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.653670 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" event={"ID":"00a5290b-dff8-438f-a047-33f1d5f05f4d","Type":"ContainerDied","Data":"11bc5859f1ffc00c8eff2f17498a0daa04b3fa60379e483d0a9c353e0fa90c70"} Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.654022 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11bc5859f1ffc00c8eff2f17498a0daa04b3fa60379e483d0a9c353e0fa90c70" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.654223 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-76smf" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.802274 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm"] Nov 22 08:36:57 crc kubenswrapper[4735]: E1122 08:36:57.803442 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a5290b-dff8-438f-a047-33f1d5f05f4d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.803519 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a5290b-dff8-438f-a047-33f1d5f05f4d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.804117 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a5290b-dff8-438f-a047-33f1d5f05f4d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.809225 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.817253 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.817330 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.818236 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.818706 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.824689 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm"] Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.859310 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.859742 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.860002 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5pwj\" (UniqueName: \"kubernetes.io/projected/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-kube-api-access-t5pwj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.961697 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5pwj\" (UniqueName: \"kubernetes.io/projected/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-kube-api-access-t5pwj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.961839 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.961945 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.967431 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.968194 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:57 crc kubenswrapper[4735]: I1122 08:36:57.979521 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5pwj\" (UniqueName: \"kubernetes.io/projected/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-kube-api-access-t5pwj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m5dfm\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:58 crc kubenswrapper[4735]: I1122 08:36:58.039324 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-mlbnx"] Nov 22 08:36:58 crc kubenswrapper[4735]: I1122 08:36:58.048628 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-mlbnx"] Nov 22 08:36:58 crc kubenswrapper[4735]: I1122 08:36:58.144527 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:36:58 crc kubenswrapper[4735]: I1122 08:36:58.754228 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm"] Nov 22 08:36:58 crc kubenswrapper[4735]: W1122 08:36:58.763088 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8e6f487_27cc_4062_8ccc_ba2a7f06454d.slice/crio-0abcffc7e235e4685bcc3a1b446ac5eb4d41408ecc7219d4db1c54660131c0ca WatchSource:0}: Error finding container 0abcffc7e235e4685bcc3a1b446ac5eb4d41408ecc7219d4db1c54660131c0ca: Status 404 returned error can't find the container with id 0abcffc7e235e4685bcc3a1b446ac5eb4d41408ecc7219d4db1c54660131c0ca Nov 22 08:36:59 crc kubenswrapper[4735]: I1122 08:36:59.284237 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f63f2a55-15ff-41f3-834f-537c0d9556fb" path="/var/lib/kubelet/pods/f63f2a55-15ff-41f3-834f-537c0d9556fb/volumes" Nov 22 08:36:59 crc kubenswrapper[4735]: I1122 08:36:59.398155 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:36:59 crc kubenswrapper[4735]: I1122 08:36:59.680302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" event={"ID":"a8e6f487-27cc-4062-8ccc-ba2a7f06454d","Type":"ContainerStarted","Data":"e7e09923fd01eb891e22b13cab4ff22461104dee05bb3fce89fc8e2e600dd6a3"} Nov 22 08:36:59 crc kubenswrapper[4735]: I1122 08:36:59.680572 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" event={"ID":"a8e6f487-27cc-4062-8ccc-ba2a7f06454d","Type":"ContainerStarted","Data":"0abcffc7e235e4685bcc3a1b446ac5eb4d41408ecc7219d4db1c54660131c0ca"} Nov 22 08:36:59 crc kubenswrapper[4735]: I1122 08:36:59.712425 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" podStartSLOduration=2.091544487 podStartE2EDuration="2.712402997s" podCreationTimestamp="2025-11-22 08:36:57 +0000 UTC" firstStartedPulling="2025-11-22 08:36:58.774125468 +0000 UTC m=+2040.378464083" lastFinishedPulling="2025-11-22 08:36:59.394983988 +0000 UTC m=+2040.999322593" observedRunningTime="2025-11-22 08:36:59.699588591 +0000 UTC m=+2041.303927196" watchObservedRunningTime="2025-11-22 08:36:59.712402997 +0000 UTC m=+2041.316741622" Nov 22 08:37:08 crc kubenswrapper[4735]: I1122 08:37:08.063237 4735 scope.go:117] "RemoveContainer" containerID="ec98ac524eab262f16419899093b6f59c27840dee024c005bfa24c59f8bd51bb" Nov 22 08:37:08 crc kubenswrapper[4735]: I1122 08:37:08.107675 4735 scope.go:117] "RemoveContainer" containerID="859217d0a75b1a510d7d855ad2228d96c4008c74831dac4111be6e84d7bbd55b" Nov 22 08:37:08 crc kubenswrapper[4735]: I1122 08:37:08.160939 4735 scope.go:117] "RemoveContainer" containerID="bb479f3f4433946fd75ead2b2be992fe15d075627bc518838bbd77131e3c843e" Nov 22 08:37:08 crc kubenswrapper[4735]: I1122 08:37:08.233586 4735 scope.go:117] "RemoveContainer" containerID="d2e9bde4bfd74a00f14234421cb0317ec1f2322ce85e0bd42c486e36ea437819" Nov 22 08:37:08 crc kubenswrapper[4735]: I1122 08:37:08.305986 4735 scope.go:117] "RemoveContainer" containerID="627f34b0078843df7c995ea540213274c097e4779b8cae6447f11f9dfc287f23" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.618875 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c47hq"] Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.623227 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.634613 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c47hq"] Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.667534 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfmvv\" (UniqueName: \"kubernetes.io/projected/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-kube-api-access-zfmvv\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.667593 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-utilities\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.667632 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-catalog-content\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.769998 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfmvv\" (UniqueName: \"kubernetes.io/projected/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-kube-api-access-zfmvv\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.770063 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-utilities\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.770105 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-catalog-content\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.770711 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-utilities\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.770772 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-catalog-content\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.793539 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfmvv\" (UniqueName: \"kubernetes.io/projected/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-kube-api-access-zfmvv\") pod \"certified-operators-c47hq\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:17 crc kubenswrapper[4735]: I1122 08:37:17.978300 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:18 crc kubenswrapper[4735]: I1122 08:37:18.554273 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c47hq"] Nov 22 08:37:18 crc kubenswrapper[4735]: I1122 08:37:18.893619 4735 generic.go:334] "Generic (PLEG): container finished" podID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerID="2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206" exitCode=0 Nov 22 08:37:18 crc kubenswrapper[4735]: I1122 08:37:18.893670 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47hq" event={"ID":"e1b798f3-f626-4a7c-a783-ca01ac45fc9d","Type":"ContainerDied","Data":"2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206"} Nov 22 08:37:18 crc kubenswrapper[4735]: I1122 08:37:18.893700 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47hq" event={"ID":"e1b798f3-f626-4a7c-a783-ca01ac45fc9d","Type":"ContainerStarted","Data":"1d0264e16bafbd0c9226bb13620abd1a8454ef3829a7a156a7335f0692a1f088"} Nov 22 08:37:19 crc kubenswrapper[4735]: I1122 08:37:19.912891 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47hq" event={"ID":"e1b798f3-f626-4a7c-a783-ca01ac45fc9d","Type":"ContainerStarted","Data":"7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6"} Nov 22 08:37:22 crc kubenswrapper[4735]: I1122 08:37:22.945748 4735 generic.go:334] "Generic (PLEG): container finished" podID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerID="7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6" exitCode=0 Nov 22 08:37:22 crc kubenswrapper[4735]: I1122 08:37:22.945827 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47hq" event={"ID":"e1b798f3-f626-4a7c-a783-ca01ac45fc9d","Type":"ContainerDied","Data":"7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6"} Nov 22 08:37:23 crc kubenswrapper[4735]: I1122 08:37:23.958618 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47hq" event={"ID":"e1b798f3-f626-4a7c-a783-ca01ac45fc9d","Type":"ContainerStarted","Data":"d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90"} Nov 22 08:37:23 crc kubenswrapper[4735]: I1122 08:37:23.987329 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c47hq" podStartSLOduration=2.496235095 podStartE2EDuration="6.987308268s" podCreationTimestamp="2025-11-22 08:37:17 +0000 UTC" firstStartedPulling="2025-11-22 08:37:18.895516806 +0000 UTC m=+2060.499855411" lastFinishedPulling="2025-11-22 08:37:23.386589959 +0000 UTC m=+2064.990928584" observedRunningTime="2025-11-22 08:37:23.977239201 +0000 UTC m=+2065.581577816" watchObservedRunningTime="2025-11-22 08:37:23.987308268 +0000 UTC m=+2065.591646883" Nov 22 08:37:27 crc kubenswrapper[4735]: I1122 08:37:27.979579 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:27 crc kubenswrapper[4735]: I1122 08:37:27.981279 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:29 crc kubenswrapper[4735]: I1122 08:37:29.034326 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-c47hq" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="registry-server" probeResult="failure" output=< Nov 22 08:37:29 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:37:29 crc kubenswrapper[4735]: > Nov 22 08:37:32 crc kubenswrapper[4735]: I1122 08:37:32.055421 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-8grl4"] Nov 22 08:37:32 crc kubenswrapper[4735]: I1122 08:37:32.069773 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-8grl4"] Nov 22 08:37:33 crc kubenswrapper[4735]: I1122 08:37:33.279728 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="272926fe-e620-4d53-b17f-524b80bdce0d" path="/var/lib/kubelet/pods/272926fe-e620-4d53-b17f-524b80bdce0d/volumes" Nov 22 08:37:38 crc kubenswrapper[4735]: I1122 08:37:38.034472 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:38 crc kubenswrapper[4735]: I1122 08:37:38.094584 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:38 crc kubenswrapper[4735]: I1122 08:37:38.274559 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c47hq"] Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.129917 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c47hq" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="registry-server" containerID="cri-o://d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90" gracePeriod=2 Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.681357 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.854362 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-utilities\") pod \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.854607 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-catalog-content\") pod \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.854678 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfmvv\" (UniqueName: \"kubernetes.io/projected/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-kube-api-access-zfmvv\") pod \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\" (UID: \"e1b798f3-f626-4a7c-a783-ca01ac45fc9d\") " Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.855681 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-utilities" (OuterVolumeSpecName: "utilities") pod "e1b798f3-f626-4a7c-a783-ca01ac45fc9d" (UID: "e1b798f3-f626-4a7c-a783-ca01ac45fc9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.863705 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-kube-api-access-zfmvv" (OuterVolumeSpecName: "kube-api-access-zfmvv") pod "e1b798f3-f626-4a7c-a783-ca01ac45fc9d" (UID: "e1b798f3-f626-4a7c-a783-ca01ac45fc9d"). InnerVolumeSpecName "kube-api-access-zfmvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.944827 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1b798f3-f626-4a7c-a783-ca01ac45fc9d" (UID: "e1b798f3-f626-4a7c-a783-ca01ac45fc9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.958454 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.958519 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:37:39 crc kubenswrapper[4735]: I1122 08:37:39.958542 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfmvv\" (UniqueName: \"kubernetes.io/projected/e1b798f3-f626-4a7c-a783-ca01ac45fc9d-kube-api-access-zfmvv\") on node \"crc\" DevicePath \"\"" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.151716 4735 generic.go:334] "Generic (PLEG): container finished" podID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerID="d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90" exitCode=0 Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.151793 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47hq" event={"ID":"e1b798f3-f626-4a7c-a783-ca01ac45fc9d","Type":"ContainerDied","Data":"d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90"} Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.152054 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47hq" event={"ID":"e1b798f3-f626-4a7c-a783-ca01ac45fc9d","Type":"ContainerDied","Data":"1d0264e16bafbd0c9226bb13620abd1a8454ef3829a7a156a7335f0692a1f088"} Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.151807 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c47hq" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.152104 4735 scope.go:117] "RemoveContainer" containerID="d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.155391 4735 generic.go:334] "Generic (PLEG): container finished" podID="a8e6f487-27cc-4062-8ccc-ba2a7f06454d" containerID="e7e09923fd01eb891e22b13cab4ff22461104dee05bb3fce89fc8e2e600dd6a3" exitCode=0 Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.155441 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" event={"ID":"a8e6f487-27cc-4062-8ccc-ba2a7f06454d","Type":"ContainerDied","Data":"e7e09923fd01eb891e22b13cab4ff22461104dee05bb3fce89fc8e2e600dd6a3"} Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.189191 4735 scope.go:117] "RemoveContainer" containerID="7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.216527 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c47hq"] Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.230100 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c47hq"] Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.233577 4735 scope.go:117] "RemoveContainer" containerID="2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.271724 4735 scope.go:117] "RemoveContainer" containerID="d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90" Nov 22 08:37:40 crc kubenswrapper[4735]: E1122 08:37:40.272293 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90\": container with ID starting with d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90 not found: ID does not exist" containerID="d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.272327 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90"} err="failed to get container status \"d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90\": rpc error: code = NotFound desc = could not find container \"d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90\": container with ID starting with d27f46399c2502f99aeb1c8dc333a3eb2a679adcb85b448d2eec66d31bf2db90 not found: ID does not exist" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.272346 4735 scope.go:117] "RemoveContainer" containerID="7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6" Nov 22 08:37:40 crc kubenswrapper[4735]: E1122 08:37:40.272846 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6\": container with ID starting with 7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6 not found: ID does not exist" containerID="7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.272880 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6"} err="failed to get container status \"7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6\": rpc error: code = NotFound desc = could not find container \"7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6\": container with ID starting with 7b182129d7ccb06a34668397e47a493cec2e0b7ee5809631fe951307308600d6 not found: ID does not exist" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.272897 4735 scope.go:117] "RemoveContainer" containerID="2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206" Nov 22 08:37:40 crc kubenswrapper[4735]: E1122 08:37:40.273128 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206\": container with ID starting with 2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206 not found: ID does not exist" containerID="2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206" Nov 22 08:37:40 crc kubenswrapper[4735]: I1122 08:37:40.273166 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206"} err="failed to get container status \"2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206\": rpc error: code = NotFound desc = could not find container \"2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206\": container with ID starting with 2378b5762e9e6c65e7d4874d998a0088a5a78726de99a209242de9daae801206 not found: ID does not exist" Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.294101 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" path="/var/lib/kubelet/pods/e1b798f3-f626-4a7c-a783-ca01ac45fc9d/volumes" Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.610257 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.800035 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-ssh-key\") pod \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.800136 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5pwj\" (UniqueName: \"kubernetes.io/projected/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-kube-api-access-t5pwj\") pod \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.800402 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-inventory\") pod \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\" (UID: \"a8e6f487-27cc-4062-8ccc-ba2a7f06454d\") " Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.805504 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-kube-api-access-t5pwj" (OuterVolumeSpecName: "kube-api-access-t5pwj") pod "a8e6f487-27cc-4062-8ccc-ba2a7f06454d" (UID: "a8e6f487-27cc-4062-8ccc-ba2a7f06454d"). InnerVolumeSpecName "kube-api-access-t5pwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.835088 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a8e6f487-27cc-4062-8ccc-ba2a7f06454d" (UID: "a8e6f487-27cc-4062-8ccc-ba2a7f06454d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.838764 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-inventory" (OuterVolumeSpecName: "inventory") pod "a8e6f487-27cc-4062-8ccc-ba2a7f06454d" (UID: "a8e6f487-27cc-4062-8ccc-ba2a7f06454d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.903390 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.903431 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:37:41 crc kubenswrapper[4735]: I1122 08:37:41.903443 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5pwj\" (UniqueName: \"kubernetes.io/projected/a8e6f487-27cc-4062-8ccc-ba2a7f06454d-kube-api-access-t5pwj\") on node \"crc\" DevicePath \"\"" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.182059 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" event={"ID":"a8e6f487-27cc-4062-8ccc-ba2a7f06454d","Type":"ContainerDied","Data":"0abcffc7e235e4685bcc3a1b446ac5eb4d41408ecc7219d4db1c54660131c0ca"} Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.182327 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0abcffc7e235e4685bcc3a1b446ac5eb4d41408ecc7219d4db1c54660131c0ca" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.182134 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m5dfm" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.314177 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx"] Nov 22 08:37:42 crc kubenswrapper[4735]: E1122 08:37:42.314750 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="extract-utilities" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.314766 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="extract-utilities" Nov 22 08:37:42 crc kubenswrapper[4735]: E1122 08:37:42.314780 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="extract-content" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.314788 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="extract-content" Nov 22 08:37:42 crc kubenswrapper[4735]: E1122 08:37:42.314825 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="registry-server" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.314834 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="registry-server" Nov 22 08:37:42 crc kubenswrapper[4735]: E1122 08:37:42.314845 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e6f487-27cc-4062-8ccc-ba2a7f06454d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.314855 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e6f487-27cc-4062-8ccc-ba2a7f06454d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.315142 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b798f3-f626-4a7c-a783-ca01ac45fc9d" containerName="registry-server" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.315157 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8e6f487-27cc-4062-8ccc-ba2a7f06454d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.316091 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.327485 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx"] Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.338973 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.339189 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.339185 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.341090 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.417701 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvv9k\" (UniqueName: \"kubernetes.io/projected/ce09bb64-050b-47f7-b2dc-2673c5444bf0-kube-api-access-wvv9k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.417822 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.418033 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.520632 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvv9k\" (UniqueName: \"kubernetes.io/projected/ce09bb64-050b-47f7-b2dc-2673c5444bf0-kube-api-access-wvv9k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.520764 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.520875 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.525779 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.526683 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.538037 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvv9k\" (UniqueName: \"kubernetes.io/projected/ce09bb64-050b-47f7-b2dc-2673c5444bf0-kube-api-access-wvv9k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:42 crc kubenswrapper[4735]: I1122 08:37:42.649639 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:37:43 crc kubenswrapper[4735]: I1122 08:37:43.287288 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx"] Nov 22 08:37:44 crc kubenswrapper[4735]: I1122 08:37:44.204935 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" event={"ID":"ce09bb64-050b-47f7-b2dc-2673c5444bf0","Type":"ContainerStarted","Data":"bb9e43e7bb5da0273f7a92374b7c7e7046d35d10a0f85dc89192fe2a861d26ba"} Nov 22 08:37:45 crc kubenswrapper[4735]: I1122 08:37:45.224050 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" event={"ID":"ce09bb64-050b-47f7-b2dc-2673c5444bf0","Type":"ContainerStarted","Data":"d9065ba1dade8fe526968e8e2cf8893ed05348829e12795ccdebfd0dbfd352f4"} Nov 22 08:37:45 crc kubenswrapper[4735]: I1122 08:37:45.264927 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" podStartSLOduration=2.5716468519999998 podStartE2EDuration="3.264902803s" podCreationTimestamp="2025-11-22 08:37:42 +0000 UTC" firstStartedPulling="2025-11-22 08:37:43.30234255 +0000 UTC m=+2084.906681155" lastFinishedPulling="2025-11-22 08:37:43.995598501 +0000 UTC m=+2085.599937106" observedRunningTime="2025-11-22 08:37:45.245132791 +0000 UTC m=+2086.849471416" watchObservedRunningTime="2025-11-22 08:37:45.264902803 +0000 UTC m=+2086.869241408" Nov 22 08:38:08 crc kubenswrapper[4735]: I1122 08:38:08.475792 4735 scope.go:117] "RemoveContainer" containerID="ae213ed10166fc9ce8fa46d4d6004f5952b5097ca6c62452515bfe43001b5df8" Nov 22 08:38:42 crc kubenswrapper[4735]: I1122 08:38:42.992091 4735 generic.go:334] "Generic (PLEG): container finished" podID="ce09bb64-050b-47f7-b2dc-2673c5444bf0" containerID="d9065ba1dade8fe526968e8e2cf8893ed05348829e12795ccdebfd0dbfd352f4" exitCode=0 Nov 22 08:38:42 crc kubenswrapper[4735]: I1122 08:38:42.992205 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" event={"ID":"ce09bb64-050b-47f7-b2dc-2673c5444bf0","Type":"ContainerDied","Data":"d9065ba1dade8fe526968e8e2cf8893ed05348829e12795ccdebfd0dbfd352f4"} Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.534052 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.651043 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-inventory\") pod \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.651166 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvv9k\" (UniqueName: \"kubernetes.io/projected/ce09bb64-050b-47f7-b2dc-2673c5444bf0-kube-api-access-wvv9k\") pod \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.651248 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-ssh-key\") pod \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\" (UID: \"ce09bb64-050b-47f7-b2dc-2673c5444bf0\") " Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.660897 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce09bb64-050b-47f7-b2dc-2673c5444bf0-kube-api-access-wvv9k" (OuterVolumeSpecName: "kube-api-access-wvv9k") pod "ce09bb64-050b-47f7-b2dc-2673c5444bf0" (UID: "ce09bb64-050b-47f7-b2dc-2673c5444bf0"). InnerVolumeSpecName "kube-api-access-wvv9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.689674 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ce09bb64-050b-47f7-b2dc-2673c5444bf0" (UID: "ce09bb64-050b-47f7-b2dc-2673c5444bf0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.709357 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-inventory" (OuterVolumeSpecName: "inventory") pod "ce09bb64-050b-47f7-b2dc-2673c5444bf0" (UID: "ce09bb64-050b-47f7-b2dc-2673c5444bf0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.766311 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvv9k\" (UniqueName: \"kubernetes.io/projected/ce09bb64-050b-47f7-b2dc-2673c5444bf0-kube-api-access-wvv9k\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.766406 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:44 crc kubenswrapper[4735]: I1122 08:38:44.766517 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce09bb64-050b-47f7-b2dc-2673c5444bf0-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.015420 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" event={"ID":"ce09bb64-050b-47f7-b2dc-2673c5444bf0","Type":"ContainerDied","Data":"bb9e43e7bb5da0273f7a92374b7c7e7046d35d10a0f85dc89192fe2a861d26ba"} Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.015480 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb9e43e7bb5da0273f7a92374b7c7e7046d35d10a0f85dc89192fe2a861d26ba" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.015535 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.112028 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-km7xk"] Nov 22 08:38:45 crc kubenswrapper[4735]: E1122 08:38:45.112764 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce09bb64-050b-47f7-b2dc-2673c5444bf0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.112793 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce09bb64-050b-47f7-b2dc-2673c5444bf0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.113106 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce09bb64-050b-47f7-b2dc-2673c5444bf0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.114197 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.117739 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.118185 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.119482 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.119548 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.143806 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-km7xk"] Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.175590 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb5m8\" (UniqueName: \"kubernetes.io/projected/8eac2dec-6b49-43a8-a603-60272938cbf4-kube-api-access-gb5m8\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.175832 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.175894 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.277711 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.277880 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb5m8\" (UniqueName: \"kubernetes.io/projected/8eac2dec-6b49-43a8-a603-60272938cbf4-kube-api-access-gb5m8\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.278157 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.284480 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.285810 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.295776 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb5m8\" (UniqueName: \"kubernetes.io/projected/8eac2dec-6b49-43a8-a603-60272938cbf4-kube-api-access-gb5m8\") pod \"ssh-known-hosts-edpm-deployment-km7xk\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:45 crc kubenswrapper[4735]: I1122 08:38:45.438073 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:46 crc kubenswrapper[4735]: W1122 08:38:46.039137 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8eac2dec_6b49_43a8_a603_60272938cbf4.slice/crio-52ed114215ed0ff4e51c01f0d0674c1b5e75c85630e2c626d4ef1e4d93c82639 WatchSource:0}: Error finding container 52ed114215ed0ff4e51c01f0d0674c1b5e75c85630e2c626d4ef1e4d93c82639: Status 404 returned error can't find the container with id 52ed114215ed0ff4e51c01f0d0674c1b5e75c85630e2c626d4ef1e4d93c82639 Nov 22 08:38:46 crc kubenswrapper[4735]: I1122 08:38:46.054611 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-km7xk"] Nov 22 08:38:47 crc kubenswrapper[4735]: I1122 08:38:47.058041 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" event={"ID":"8eac2dec-6b49-43a8-a603-60272938cbf4","Type":"ContainerStarted","Data":"2a87e8c39bcf9511cb3d3ad54fcbc1a7663558f4b12961912b50b57e46321726"} Nov 22 08:38:47 crc kubenswrapper[4735]: I1122 08:38:47.058550 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" event={"ID":"8eac2dec-6b49-43a8-a603-60272938cbf4","Type":"ContainerStarted","Data":"52ed114215ed0ff4e51c01f0d0674c1b5e75c85630e2c626d4ef1e4d93c82639"} Nov 22 08:38:47 crc kubenswrapper[4735]: I1122 08:38:47.077645 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" podStartSLOduration=1.63114643 podStartE2EDuration="2.077623159s" podCreationTimestamp="2025-11-22 08:38:45 +0000 UTC" firstStartedPulling="2025-11-22 08:38:46.041915363 +0000 UTC m=+2147.646253968" lastFinishedPulling="2025-11-22 08:38:46.488392092 +0000 UTC m=+2148.092730697" observedRunningTime="2025-11-22 08:38:47.069980934 +0000 UTC m=+2148.674319559" watchObservedRunningTime="2025-11-22 08:38:47.077623159 +0000 UTC m=+2148.681961764" Nov 22 08:38:54 crc kubenswrapper[4735]: I1122 08:38:54.133550 4735 generic.go:334] "Generic (PLEG): container finished" podID="8eac2dec-6b49-43a8-a603-60272938cbf4" containerID="2a87e8c39bcf9511cb3d3ad54fcbc1a7663558f4b12961912b50b57e46321726" exitCode=0 Nov 22 08:38:54 crc kubenswrapper[4735]: I1122 08:38:54.133666 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" event={"ID":"8eac2dec-6b49-43a8-a603-60272938cbf4","Type":"ContainerDied","Data":"2a87e8c39bcf9511cb3d3ad54fcbc1a7663558f4b12961912b50b57e46321726"} Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.628108 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.807078 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-ssh-key-openstack-edpm-ipam\") pod \"8eac2dec-6b49-43a8-a603-60272938cbf4\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.807278 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-inventory-0\") pod \"8eac2dec-6b49-43a8-a603-60272938cbf4\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.807362 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb5m8\" (UniqueName: \"kubernetes.io/projected/8eac2dec-6b49-43a8-a603-60272938cbf4-kube-api-access-gb5m8\") pod \"8eac2dec-6b49-43a8-a603-60272938cbf4\" (UID: \"8eac2dec-6b49-43a8-a603-60272938cbf4\") " Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.815681 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eac2dec-6b49-43a8-a603-60272938cbf4-kube-api-access-gb5m8" (OuterVolumeSpecName: "kube-api-access-gb5m8") pod "8eac2dec-6b49-43a8-a603-60272938cbf4" (UID: "8eac2dec-6b49-43a8-a603-60272938cbf4"). InnerVolumeSpecName "kube-api-access-gb5m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.844498 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8eac2dec-6b49-43a8-a603-60272938cbf4" (UID: "8eac2dec-6b49-43a8-a603-60272938cbf4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.884642 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "8eac2dec-6b49-43a8-a603-60272938cbf4" (UID: "8eac2dec-6b49-43a8-a603-60272938cbf4"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.910291 4735 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.910329 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb5m8\" (UniqueName: \"kubernetes.io/projected/8eac2dec-6b49-43a8-a603-60272938cbf4-kube-api-access-gb5m8\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:55 crc kubenswrapper[4735]: I1122 08:38:55.910340 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8eac2dec-6b49-43a8-a603-60272938cbf4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.159972 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" event={"ID":"8eac2dec-6b49-43a8-a603-60272938cbf4","Type":"ContainerDied","Data":"52ed114215ed0ff4e51c01f0d0674c1b5e75c85630e2c626d4ef1e4d93c82639"} Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.160021 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52ed114215ed0ff4e51c01f0d0674c1b5e75c85630e2c626d4ef1e4d93c82639" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.160024 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-km7xk" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.232407 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6"] Nov 22 08:38:56 crc kubenswrapper[4735]: E1122 08:38:56.233000 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eac2dec-6b49-43a8-a603-60272938cbf4" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.233020 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eac2dec-6b49-43a8-a603-60272938cbf4" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.233359 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eac2dec-6b49-43a8-a603-60272938cbf4" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.234252 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.237268 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.237370 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.239071 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.239266 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.241441 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6"] Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.421419 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.421762 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.421802 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nhzb\" (UniqueName: \"kubernetes.io/projected/179b352a-82a0-438e-9e6d-79fa7b20ac90-kube-api-access-7nhzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.523396 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.523445 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nhzb\" (UniqueName: \"kubernetes.io/projected/179b352a-82a0-438e-9e6d-79fa7b20ac90-kube-api-access-7nhzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.523514 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.528149 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.528324 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.541162 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nhzb\" (UniqueName: \"kubernetes.io/projected/179b352a-82a0-438e-9e6d-79fa7b20ac90-kube-api-access-7nhzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6vlw6\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:56 crc kubenswrapper[4735]: I1122 08:38:56.593654 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:38:57 crc kubenswrapper[4735]: I1122 08:38:57.156247 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6"] Nov 22 08:38:57 crc kubenswrapper[4735]: I1122 08:38:57.171717 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" event={"ID":"179b352a-82a0-438e-9e6d-79fa7b20ac90","Type":"ContainerStarted","Data":"6617e4bde60e24f9ac0946b6d7a5fdfcedbde0e5a11c5a3a542c99f212b36084"} Nov 22 08:38:58 crc kubenswrapper[4735]: I1122 08:38:58.182907 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" event={"ID":"179b352a-82a0-438e-9e6d-79fa7b20ac90","Type":"ContainerStarted","Data":"a64edbd9926abb828cc9b893b33b11dde63e20637a480f55322da21c86250742"} Nov 22 08:38:58 crc kubenswrapper[4735]: I1122 08:38:58.208163 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" podStartSLOduration=1.780385458 podStartE2EDuration="2.20813707s" podCreationTimestamp="2025-11-22 08:38:56 +0000 UTC" firstStartedPulling="2025-11-22 08:38:57.160003649 +0000 UTC m=+2158.764342254" lastFinishedPulling="2025-11-22 08:38:57.587755261 +0000 UTC m=+2159.192093866" observedRunningTime="2025-11-22 08:38:58.202150778 +0000 UTC m=+2159.806489393" watchObservedRunningTime="2025-11-22 08:38:58.20813707 +0000 UTC m=+2159.812475675" Nov 22 08:39:03 crc kubenswrapper[4735]: I1122 08:39:03.971867 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x5g97"] Nov 22 08:39:03 crc kubenswrapper[4735]: I1122 08:39:03.974630 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:03 crc kubenswrapper[4735]: I1122 08:39:03.992141 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5g97"] Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.117345 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knpg9\" (UniqueName: \"kubernetes.io/projected/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-kube-api-access-knpg9\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.117491 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-utilities\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.117569 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-catalog-content\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.220084 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-utilities\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.220224 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-catalog-content\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.220340 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knpg9\" (UniqueName: \"kubernetes.io/projected/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-kube-api-access-knpg9\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.220732 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-utilities\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.220797 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-catalog-content\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.241355 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knpg9\" (UniqueName: \"kubernetes.io/projected/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-kube-api-access-knpg9\") pod \"redhat-marketplace-x5g97\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.303830 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:04 crc kubenswrapper[4735]: I1122 08:39:04.892274 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5g97"] Nov 22 08:39:05 crc kubenswrapper[4735]: I1122 08:39:05.268530 4735 generic.go:334] "Generic (PLEG): container finished" podID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerID="28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f" exitCode=0 Nov 22 08:39:05 crc kubenswrapper[4735]: I1122 08:39:05.280572 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5g97" event={"ID":"c5ef2575-9b67-48cd-8dc5-83b2db5804fd","Type":"ContainerDied","Data":"28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f"} Nov 22 08:39:05 crc kubenswrapper[4735]: I1122 08:39:05.280624 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5g97" event={"ID":"c5ef2575-9b67-48cd-8dc5-83b2db5804fd","Type":"ContainerStarted","Data":"61aade8af810e1e77ec9ceb762615f09573f6831a6defc023fbd3b4551249942"} Nov 22 08:39:06 crc kubenswrapper[4735]: I1122 08:39:06.281816 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5g97" event={"ID":"c5ef2575-9b67-48cd-8dc5-83b2db5804fd","Type":"ContainerStarted","Data":"46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947"} Nov 22 08:39:07 crc kubenswrapper[4735]: I1122 08:39:07.297707 4735 generic.go:334] "Generic (PLEG): container finished" podID="179b352a-82a0-438e-9e6d-79fa7b20ac90" containerID="a64edbd9926abb828cc9b893b33b11dde63e20637a480f55322da21c86250742" exitCode=0 Nov 22 08:39:07 crc kubenswrapper[4735]: I1122 08:39:07.297820 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" event={"ID":"179b352a-82a0-438e-9e6d-79fa7b20ac90","Type":"ContainerDied","Data":"a64edbd9926abb828cc9b893b33b11dde63e20637a480f55322da21c86250742"} Nov 22 08:39:07 crc kubenswrapper[4735]: I1122 08:39:07.301178 4735 generic.go:334] "Generic (PLEG): container finished" podID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerID="46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947" exitCode=0 Nov 22 08:39:07 crc kubenswrapper[4735]: I1122 08:39:07.301220 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5g97" event={"ID":"c5ef2575-9b67-48cd-8dc5-83b2db5804fd","Type":"ContainerDied","Data":"46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947"} Nov 22 08:39:08 crc kubenswrapper[4735]: I1122 08:39:08.319748 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5g97" event={"ID":"c5ef2575-9b67-48cd-8dc5-83b2db5804fd","Type":"ContainerStarted","Data":"e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730"} Nov 22 08:39:08 crc kubenswrapper[4735]: I1122 08:39:08.352104 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x5g97" podStartSLOduration=2.892140329 podStartE2EDuration="5.352085015s" podCreationTimestamp="2025-11-22 08:39:03 +0000 UTC" firstStartedPulling="2025-11-22 08:39:05.27176105 +0000 UTC m=+2166.876099655" lastFinishedPulling="2025-11-22 08:39:07.731705746 +0000 UTC m=+2169.336044341" observedRunningTime="2025-11-22 08:39:08.341918596 +0000 UTC m=+2169.946257201" watchObservedRunningTime="2025-11-22 08:39:08.352085015 +0000 UTC m=+2169.956423620" Nov 22 08:39:08 crc kubenswrapper[4735]: I1122 08:39:08.884046 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.038558 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-inventory\") pod \"179b352a-82a0-438e-9e6d-79fa7b20ac90\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.039673 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nhzb\" (UniqueName: \"kubernetes.io/projected/179b352a-82a0-438e-9e6d-79fa7b20ac90-kube-api-access-7nhzb\") pod \"179b352a-82a0-438e-9e6d-79fa7b20ac90\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.039719 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-ssh-key\") pod \"179b352a-82a0-438e-9e6d-79fa7b20ac90\" (UID: \"179b352a-82a0-438e-9e6d-79fa7b20ac90\") " Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.044877 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/179b352a-82a0-438e-9e6d-79fa7b20ac90-kube-api-access-7nhzb" (OuterVolumeSpecName: "kube-api-access-7nhzb") pod "179b352a-82a0-438e-9e6d-79fa7b20ac90" (UID: "179b352a-82a0-438e-9e6d-79fa7b20ac90"). InnerVolumeSpecName "kube-api-access-7nhzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.077596 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-inventory" (OuterVolumeSpecName: "inventory") pod "179b352a-82a0-438e-9e6d-79fa7b20ac90" (UID: "179b352a-82a0-438e-9e6d-79fa7b20ac90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.085664 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "179b352a-82a0-438e-9e6d-79fa7b20ac90" (UID: "179b352a-82a0-438e-9e6d-79fa7b20ac90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.144009 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.144061 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nhzb\" (UniqueName: \"kubernetes.io/projected/179b352a-82a0-438e-9e6d-79fa7b20ac90-kube-api-access-7nhzb\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.144085 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/179b352a-82a0-438e-9e6d-79fa7b20ac90-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.332133 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.332115 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6vlw6" event={"ID":"179b352a-82a0-438e-9e6d-79fa7b20ac90","Type":"ContainerDied","Data":"6617e4bde60e24f9ac0946b6d7a5fdfcedbde0e5a11c5a3a542c99f212b36084"} Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.333407 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6617e4bde60e24f9ac0946b6d7a5fdfcedbde0e5a11c5a3a542c99f212b36084" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.402331 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw"] Nov 22 08:39:09 crc kubenswrapper[4735]: E1122 08:39:09.403166 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179b352a-82a0-438e-9e6d-79fa7b20ac90" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.403199 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="179b352a-82a0-438e-9e6d-79fa7b20ac90" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.403629 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="179b352a-82a0-438e-9e6d-79fa7b20ac90" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.405025 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.407745 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.408550 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.409088 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.410753 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.421512 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw"] Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.553387 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89pgl\" (UniqueName: \"kubernetes.io/projected/99efb16a-02eb-434a-9928-7c782fd7fc86-kube-api-access-89pgl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.553476 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.553569 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.655647 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.655909 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89pgl\" (UniqueName: \"kubernetes.io/projected/99efb16a-02eb-434a-9928-7c782fd7fc86-kube-api-access-89pgl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.656002 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.661560 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.662988 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.677890 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89pgl\" (UniqueName: \"kubernetes.io/projected/99efb16a-02eb-434a-9928-7c782fd7fc86-kube-api-access-89pgl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:09 crc kubenswrapper[4735]: I1122 08:39:09.754997 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:10 crc kubenswrapper[4735]: I1122 08:39:10.375682 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw"] Nov 22 08:39:11 crc kubenswrapper[4735]: I1122 08:39:11.357022 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" event={"ID":"99efb16a-02eb-434a-9928-7c782fd7fc86","Type":"ContainerStarted","Data":"64ad9735be50bb8c4febe05b01c47c605362d41f09267137de63981d7700117f"} Nov 22 08:39:12 crc kubenswrapper[4735]: I1122 08:39:12.384841 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" event={"ID":"99efb16a-02eb-434a-9928-7c782fd7fc86","Type":"ContainerStarted","Data":"7b9c5b684ecc9c277a38f2864f1898c8139ab98d0d265f00d211ce6d4a0b875f"} Nov 22 08:39:12 crc kubenswrapper[4735]: I1122 08:39:12.410576 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" podStartSLOduration=2.6819565340000002 podStartE2EDuration="3.410548983s" podCreationTimestamp="2025-11-22 08:39:09 +0000 UTC" firstStartedPulling="2025-11-22 08:39:10.379893478 +0000 UTC m=+2171.984232083" lastFinishedPulling="2025-11-22 08:39:11.108485897 +0000 UTC m=+2172.712824532" observedRunningTime="2025-11-22 08:39:12.40102356 +0000 UTC m=+2174.005362175" watchObservedRunningTime="2025-11-22 08:39:12.410548983 +0000 UTC m=+2174.014887598" Nov 22 08:39:14 crc kubenswrapper[4735]: I1122 08:39:14.304848 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:14 crc kubenswrapper[4735]: I1122 08:39:14.305152 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:14 crc kubenswrapper[4735]: I1122 08:39:14.362091 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:14 crc kubenswrapper[4735]: I1122 08:39:14.462543 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:14 crc kubenswrapper[4735]: I1122 08:39:14.611074 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5g97"] Nov 22 08:39:16 crc kubenswrapper[4735]: I1122 08:39:16.131950 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:39:16 crc kubenswrapper[4735]: I1122 08:39:16.132556 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:39:16 crc kubenswrapper[4735]: I1122 08:39:16.436785 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x5g97" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerName="registry-server" containerID="cri-o://e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730" gracePeriod=2 Nov 22 08:39:16 crc kubenswrapper[4735]: I1122 08:39:16.958546 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.060764 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knpg9\" (UniqueName: \"kubernetes.io/projected/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-kube-api-access-knpg9\") pod \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.061528 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-catalog-content\") pod \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.062408 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-utilities\") pod \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\" (UID: \"c5ef2575-9b67-48cd-8dc5-83b2db5804fd\") " Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.063292 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-utilities" (OuterVolumeSpecName: "utilities") pod "c5ef2575-9b67-48cd-8dc5-83b2db5804fd" (UID: "c5ef2575-9b67-48cd-8dc5-83b2db5804fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.065261 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.066562 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-kube-api-access-knpg9" (OuterVolumeSpecName: "kube-api-access-knpg9") pod "c5ef2575-9b67-48cd-8dc5-83b2db5804fd" (UID: "c5ef2575-9b67-48cd-8dc5-83b2db5804fd"). InnerVolumeSpecName "kube-api-access-knpg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.167245 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knpg9\" (UniqueName: \"kubernetes.io/projected/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-kube-api-access-knpg9\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.448786 4735 generic.go:334] "Generic (PLEG): container finished" podID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerID="e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730" exitCode=0 Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.448827 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5g97" event={"ID":"c5ef2575-9b67-48cd-8dc5-83b2db5804fd","Type":"ContainerDied","Data":"e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730"} Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.448848 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5g97" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.448860 4735 scope.go:117] "RemoveContainer" containerID="e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.448851 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5g97" event={"ID":"c5ef2575-9b67-48cd-8dc5-83b2db5804fd","Type":"ContainerDied","Data":"61aade8af810e1e77ec9ceb762615f09573f6831a6defc023fbd3b4551249942"} Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.481171 4735 scope.go:117] "RemoveContainer" containerID="46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.508285 4735 scope.go:117] "RemoveContainer" containerID="28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.591828 4735 scope.go:117] "RemoveContainer" containerID="e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730" Nov 22 08:39:17 crc kubenswrapper[4735]: E1122 08:39:17.592223 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730\": container with ID starting with e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730 not found: ID does not exist" containerID="e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.592277 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730"} err="failed to get container status \"e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730\": rpc error: code = NotFound desc = could not find container \"e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730\": container with ID starting with e948048e35cf872bc99bbc41fa88a1c15fa033a2aab4e0c0c1ed80ee607e1730 not found: ID does not exist" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.592310 4735 scope.go:117] "RemoveContainer" containerID="46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947" Nov 22 08:39:17 crc kubenswrapper[4735]: E1122 08:39:17.592936 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947\": container with ID starting with 46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947 not found: ID does not exist" containerID="46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.592981 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947"} err="failed to get container status \"46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947\": rpc error: code = NotFound desc = could not find container \"46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947\": container with ID starting with 46309b727a98736a615ad424dd2d058b9385da6ccd223a8bc3cb9cc00804b947 not found: ID does not exist" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.593010 4735 scope.go:117] "RemoveContainer" containerID="28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f" Nov 22 08:39:17 crc kubenswrapper[4735]: E1122 08:39:17.593295 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f\": container with ID starting with 28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f not found: ID does not exist" containerID="28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.593335 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f"} err="failed to get container status \"28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f\": rpc error: code = NotFound desc = could not find container \"28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f\": container with ID starting with 28692526c7815cab56f177b640889787c4ff53be720dbb51d52bef813d7ae31f not found: ID does not exist" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.656986 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5ef2575-9b67-48cd-8dc5-83b2db5804fd" (UID: "c5ef2575-9b67-48cd-8dc5-83b2db5804fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.679781 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef2575-9b67-48cd-8dc5-83b2db5804fd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.798407 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5g97"] Nov 22 08:39:17 crc kubenswrapper[4735]: I1122 08:39:17.810121 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5g97"] Nov 22 08:39:19 crc kubenswrapper[4735]: I1122 08:39:19.280325 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" path="/var/lib/kubelet/pods/c5ef2575-9b67-48cd-8dc5-83b2db5804fd/volumes" Nov 22 08:39:22 crc kubenswrapper[4735]: E1122 08:39:22.468378 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99efb16a_02eb_434a_9928_7c782fd7fc86.slice/crio-conmon-7b9c5b684ecc9c277a38f2864f1898c8139ab98d0d265f00d211ce6d4a0b875f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99efb16a_02eb_434a_9928_7c782fd7fc86.slice/crio-7b9c5b684ecc9c277a38f2864f1898c8139ab98d0d265f00d211ce6d4a0b875f.scope\": RecentStats: unable to find data in memory cache]" Nov 22 08:39:22 crc kubenswrapper[4735]: I1122 08:39:22.516303 4735 generic.go:334] "Generic (PLEG): container finished" podID="99efb16a-02eb-434a-9928-7c782fd7fc86" containerID="7b9c5b684ecc9c277a38f2864f1898c8139ab98d0d265f00d211ce6d4a0b875f" exitCode=0 Nov 22 08:39:22 crc kubenswrapper[4735]: I1122 08:39:22.516447 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" event={"ID":"99efb16a-02eb-434a-9928-7c782fd7fc86","Type":"ContainerDied","Data":"7b9c5b684ecc9c277a38f2864f1898c8139ab98d0d265f00d211ce6d4a0b875f"} Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.092529 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.146843 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-ssh-key\") pod \"99efb16a-02eb-434a-9928-7c782fd7fc86\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.147246 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-inventory\") pod \"99efb16a-02eb-434a-9928-7c782fd7fc86\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.147389 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89pgl\" (UniqueName: \"kubernetes.io/projected/99efb16a-02eb-434a-9928-7c782fd7fc86-kube-api-access-89pgl\") pod \"99efb16a-02eb-434a-9928-7c782fd7fc86\" (UID: \"99efb16a-02eb-434a-9928-7c782fd7fc86\") " Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.158179 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99efb16a-02eb-434a-9928-7c782fd7fc86-kube-api-access-89pgl" (OuterVolumeSpecName: "kube-api-access-89pgl") pod "99efb16a-02eb-434a-9928-7c782fd7fc86" (UID: "99efb16a-02eb-434a-9928-7c782fd7fc86"). InnerVolumeSpecName "kube-api-access-89pgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.203811 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-inventory" (OuterVolumeSpecName: "inventory") pod "99efb16a-02eb-434a-9928-7c782fd7fc86" (UID: "99efb16a-02eb-434a-9928-7c782fd7fc86"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.209631 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "99efb16a-02eb-434a-9928-7c782fd7fc86" (UID: "99efb16a-02eb-434a-9928-7c782fd7fc86"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.251704 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89pgl\" (UniqueName: \"kubernetes.io/projected/99efb16a-02eb-434a-9928-7c782fd7fc86-kube-api-access-89pgl\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.251756 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.251768 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99efb16a-02eb-434a-9928-7c782fd7fc86-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.543292 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" event={"ID":"99efb16a-02eb-434a-9928-7c782fd7fc86","Type":"ContainerDied","Data":"64ad9735be50bb8c4febe05b01c47c605362d41f09267137de63981d7700117f"} Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.543742 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64ad9735be50bb8c4febe05b01c47c605362d41f09267137de63981d7700117f" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.543577 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.638184 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z"] Nov 22 08:39:24 crc kubenswrapper[4735]: E1122 08:39:24.638665 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99efb16a-02eb-434a-9928-7c782fd7fc86" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.638687 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="99efb16a-02eb-434a-9928-7c782fd7fc86" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:24 crc kubenswrapper[4735]: E1122 08:39:24.638706 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerName="extract-utilities" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.638714 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerName="extract-utilities" Nov 22 08:39:24 crc kubenswrapper[4735]: E1122 08:39:24.638755 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerName="registry-server" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.638764 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerName="registry-server" Nov 22 08:39:24 crc kubenswrapper[4735]: E1122 08:39:24.638792 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerName="extract-content" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.638800 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerName="extract-content" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.639096 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="99efb16a-02eb-434a-9928-7c782fd7fc86" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.639130 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ef2575-9b67-48cd-8dc5-83b2db5804fd" containerName="registry-server" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.639971 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.642719 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.643164 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.643242 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.643249 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.643427 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.643436 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.643741 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.643921 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.644490 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.669752 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z"] Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.766856 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.766959 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.767242 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.767307 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.767351 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.767448 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.767563 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.767630 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2lvp\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-kube-api-access-j2lvp\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.767710 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.768109 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.768166 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.768323 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.768420 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.768484 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.768545 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.768591 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.870868 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871004 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871060 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871102 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871136 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871161 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871200 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871234 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871269 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871311 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871368 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871401 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871427 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871590 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871639 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.871678 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2lvp\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-kube-api-access-j2lvp\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.876441 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.878084 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.879033 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.879308 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.879583 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.880067 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.880273 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.880350 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.881119 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.881717 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.882421 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.883978 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.884355 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.884546 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.885872 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.901097 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2lvp\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-kube-api-access-j2lvp\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kth9z\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:24 crc kubenswrapper[4735]: I1122 08:39:24.972648 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:39:25 crc kubenswrapper[4735]: I1122 08:39:25.608411 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z"] Nov 22 08:39:25 crc kubenswrapper[4735]: W1122 08:39:25.613655 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod506cbc0b_4b85_42e9_a73a_52721f001c05.slice/crio-ad32e0c9ad2711af7860dc21e2516d0c99419345b88eea29da9de03bf19876f4 WatchSource:0}: Error finding container ad32e0c9ad2711af7860dc21e2516d0c99419345b88eea29da9de03bf19876f4: Status 404 returned error can't find the container with id ad32e0c9ad2711af7860dc21e2516d0c99419345b88eea29da9de03bf19876f4 Nov 22 08:39:26 crc kubenswrapper[4735]: I1122 08:39:26.565005 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" event={"ID":"506cbc0b-4b85-42e9-a73a-52721f001c05","Type":"ContainerStarted","Data":"dff9ed09be2f4e835199a81a2a45a067bc67fa660ddf304aa7a6b66220a0218f"} Nov 22 08:39:26 crc kubenswrapper[4735]: I1122 08:39:26.565517 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" event={"ID":"506cbc0b-4b85-42e9-a73a-52721f001c05","Type":"ContainerStarted","Data":"ad32e0c9ad2711af7860dc21e2516d0c99419345b88eea29da9de03bf19876f4"} Nov 22 08:39:26 crc kubenswrapper[4735]: I1122 08:39:26.598043 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" podStartSLOduration=2.184251627 podStartE2EDuration="2.598014774s" podCreationTimestamp="2025-11-22 08:39:24 +0000 UTC" firstStartedPulling="2025-11-22 08:39:25.615940202 +0000 UTC m=+2187.220278807" lastFinishedPulling="2025-11-22 08:39:26.029703349 +0000 UTC m=+2187.634041954" observedRunningTime="2025-11-22 08:39:26.587526487 +0000 UTC m=+2188.191865142" watchObservedRunningTime="2025-11-22 08:39:26.598014774 +0000 UTC m=+2188.202353399" Nov 22 08:39:38 crc kubenswrapper[4735]: I1122 08:39:38.052746 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-7wl2v"] Nov 22 08:39:38 crc kubenswrapper[4735]: I1122 08:39:38.065289 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-7wl2v"] Nov 22 08:39:39 crc kubenswrapper[4735]: I1122 08:39:39.284162 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca27f7d6-a068-47bb-a286-1941df09fdec" path="/var/lib/kubelet/pods/ca27f7d6-a068-47bb-a286-1941df09fdec/volumes" Nov 22 08:39:46 crc kubenswrapper[4735]: I1122 08:39:46.131746 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:39:46 crc kubenswrapper[4735]: I1122 08:39:46.132607 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:40:08 crc kubenswrapper[4735]: I1122 08:40:08.591650 4735 scope.go:117] "RemoveContainer" containerID="3d590a4899629b1574a875232b61d00da1a8bcceb36ce04219789036ce50496f" Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.131230 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.131968 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.132036 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.133545 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.133660 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" gracePeriod=600 Nov 22 08:40:16 crc kubenswrapper[4735]: E1122 08:40:16.267369 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.372031 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" exitCode=0 Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.372117 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9"} Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.372164 4735 scope.go:117] "RemoveContainer" containerID="4681e22fbcdfadcc11184336c0fc4ee081ea57446a881d6cbac7c3aa51dc6f7c" Nov 22 08:40:16 crc kubenswrapper[4735]: I1122 08:40:16.374203 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:40:16 crc kubenswrapper[4735]: E1122 08:40:16.376249 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:40:19 crc kubenswrapper[4735]: I1122 08:40:19.418867 4735 generic.go:334] "Generic (PLEG): container finished" podID="506cbc0b-4b85-42e9-a73a-52721f001c05" containerID="dff9ed09be2f4e835199a81a2a45a067bc67fa660ddf304aa7a6b66220a0218f" exitCode=0 Nov 22 08:40:19 crc kubenswrapper[4735]: I1122 08:40:19.419930 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" event={"ID":"506cbc0b-4b85-42e9-a73a-52721f001c05","Type":"ContainerDied","Data":"dff9ed09be2f4e835199a81a2a45a067bc67fa660ddf304aa7a6b66220a0218f"} Nov 22 08:40:20 crc kubenswrapper[4735]: I1122 08:40:20.072060 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-74rcz"] Nov 22 08:40:20 crc kubenswrapper[4735]: I1122 08:40:20.086795 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-74rcz"] Nov 22 08:40:20 crc kubenswrapper[4735]: I1122 08:40:20.963904 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.135006 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.135321 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-power-monitoring-combined-ca-bundle\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136230 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136265 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-combined-ca-bundle\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136324 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ssh-key\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136416 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-ovn-default-certs-0\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136522 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2lvp\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-kube-api-access-j2lvp\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136567 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-neutron-metadata-combined-ca-bundle\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136612 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-repo-setup-combined-ca-bundle\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136643 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-libvirt-combined-ca-bundle\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136691 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136761 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136925 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ovn-combined-ca-bundle\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.136985 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-inventory\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.137026 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-bootstrap-combined-ca-bundle\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.137093 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-nova-combined-ca-bundle\") pod \"506cbc0b-4b85-42e9-a73a-52721f001c05\" (UID: \"506cbc0b-4b85-42e9-a73a-52721f001c05\") " Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.142554 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-kube-api-access-j2lvp" (OuterVolumeSpecName: "kube-api-access-j2lvp") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "kube-api-access-j2lvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.143258 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.144112 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.146271 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.146967 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.149832 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.149999 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.150223 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.150302 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.150603 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.150688 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.151001 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.151568 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.153767 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.177921 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-inventory" (OuterVolumeSpecName: "inventory") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.182969 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "506cbc0b-4b85-42e9-a73a-52721f001c05" (UID: "506cbc0b-4b85-42e9-a73a-52721f001c05"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240250 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240289 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2lvp\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-kube-api-access-j2lvp\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240301 4735 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240311 4735 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240323 4735 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240353 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240365 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240376 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240386 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240394 4735 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240402 4735 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240411 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240423 4735 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240434 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/506cbc0b-4b85-42e9-a73a-52721f001c05-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240443 4735 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.240465 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/506cbc0b-4b85-42e9-a73a-52721f001c05-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.280622 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b3826dc-bab1-4c6d-85dc-cecf267d0c70" path="/var/lib/kubelet/pods/5b3826dc-bab1-4c6d-85dc-cecf267d0c70/volumes" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.450254 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" event={"ID":"506cbc0b-4b85-42e9-a73a-52721f001c05","Type":"ContainerDied","Data":"ad32e0c9ad2711af7860dc21e2516d0c99419345b88eea29da9de03bf19876f4"} Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.450634 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad32e0c9ad2711af7860dc21e2516d0c99419345b88eea29da9de03bf19876f4" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.450335 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kth9z" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.571969 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6"] Nov 22 08:40:21 crc kubenswrapper[4735]: E1122 08:40:21.572540 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="506cbc0b-4b85-42e9-a73a-52721f001c05" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.572564 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="506cbc0b-4b85-42e9-a73a-52721f001c05" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.572883 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="506cbc0b-4b85-42e9-a73a-52721f001c05" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.573947 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.579168 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.579174 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.579298 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.579184 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.579386 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.601565 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6"] Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.753851 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/311958ee-5838-4326-9b40-7b7228bd6393-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.754274 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.754324 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4txwd\" (UniqueName: \"kubernetes.io/projected/311958ee-5838-4326-9b40-7b7228bd6393-kube-api-access-4txwd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.754445 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.754510 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.855922 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.855990 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4txwd\" (UniqueName: \"kubernetes.io/projected/311958ee-5838-4326-9b40-7b7228bd6393-kube-api-access-4txwd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.856078 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.856099 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.856151 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/311958ee-5838-4326-9b40-7b7228bd6393-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.857091 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/311958ee-5838-4326-9b40-7b7228bd6393-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.860097 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.861238 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.862732 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.879643 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4txwd\" (UniqueName: \"kubernetes.io/projected/311958ee-5838-4326-9b40-7b7228bd6393-kube-api-access-4txwd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lgnq6\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:21 crc kubenswrapper[4735]: I1122 08:40:21.892944 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:40:22 crc kubenswrapper[4735]: I1122 08:40:22.505803 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6"] Nov 22 08:40:22 crc kubenswrapper[4735]: W1122 08:40:22.515337 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod311958ee_5838_4326_9b40_7b7228bd6393.slice/crio-006cbdb1c894eaf7638bf3276af1e55d505d1328fb4350c5d85ffbcddd2fcd6c WatchSource:0}: Error finding container 006cbdb1c894eaf7638bf3276af1e55d505d1328fb4350c5d85ffbcddd2fcd6c: Status 404 returned error can't find the container with id 006cbdb1c894eaf7638bf3276af1e55d505d1328fb4350c5d85ffbcddd2fcd6c Nov 22 08:40:23 crc kubenswrapper[4735]: I1122 08:40:23.481403 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" event={"ID":"311958ee-5838-4326-9b40-7b7228bd6393","Type":"ContainerStarted","Data":"d0ddda72651d8d2964a566ec27b26c79eabf2591dad5ea4a74fe71dc813dd954"} Nov 22 08:40:23 crc kubenswrapper[4735]: I1122 08:40:23.482450 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" event={"ID":"311958ee-5838-4326-9b40-7b7228bd6393","Type":"ContainerStarted","Data":"006cbdb1c894eaf7638bf3276af1e55d505d1328fb4350c5d85ffbcddd2fcd6c"} Nov 22 08:40:23 crc kubenswrapper[4735]: I1122 08:40:23.522365 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" podStartSLOduration=2.081188884 podStartE2EDuration="2.51893083s" podCreationTimestamp="2025-11-22 08:40:21 +0000 UTC" firstStartedPulling="2025-11-22 08:40:22.519624186 +0000 UTC m=+2244.123962821" lastFinishedPulling="2025-11-22 08:40:22.957366152 +0000 UTC m=+2244.561704767" observedRunningTime="2025-11-22 08:40:23.512638507 +0000 UTC m=+2245.116977142" watchObservedRunningTime="2025-11-22 08:40:23.51893083 +0000 UTC m=+2245.123269465" Nov 22 08:40:29 crc kubenswrapper[4735]: I1122 08:40:29.276213 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:40:29 crc kubenswrapper[4735]: E1122 08:40:29.277285 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:40:44 crc kubenswrapper[4735]: I1122 08:40:44.263715 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:40:44 crc kubenswrapper[4735]: E1122 08:40:44.264347 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:40:56 crc kubenswrapper[4735]: I1122 08:40:56.263647 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:40:56 crc kubenswrapper[4735]: E1122 08:40:56.264769 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:41:08 crc kubenswrapper[4735]: I1122 08:41:08.702257 4735 scope.go:117] "RemoveContainer" containerID="f0a4c584a0695c3ad4cf7fb3298c1047dc462aa6264a4cf143174834fa59e71a" Nov 22 08:41:10 crc kubenswrapper[4735]: I1122 08:41:10.264384 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:41:10 crc kubenswrapper[4735]: E1122 08:41:10.265198 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.263788 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:41:25 crc kubenswrapper[4735]: E1122 08:41:25.264756 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.503659 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8hd8r"] Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.508149 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.515972 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8hd8r"] Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.604723 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-utilities\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.604784 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx6sb\" (UniqueName: \"kubernetes.io/projected/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-kube-api-access-sx6sb\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.605279 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-catalog-content\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.710218 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-utilities\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.710271 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx6sb\" (UniqueName: \"kubernetes.io/projected/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-kube-api-access-sx6sb\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.710330 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-catalog-content\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.710835 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-utilities\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.710874 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-catalog-content\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.735377 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx6sb\" (UniqueName: \"kubernetes.io/projected/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-kube-api-access-sx6sb\") pod \"redhat-operators-8hd8r\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:25 crc kubenswrapper[4735]: I1122 08:41:25.832443 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:26 crc kubenswrapper[4735]: I1122 08:41:26.350647 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8hd8r"] Nov 22 08:41:27 crc kubenswrapper[4735]: I1122 08:41:27.269817 4735 generic.go:334] "Generic (PLEG): container finished" podID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerID="24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26" exitCode=0 Nov 22 08:41:27 crc kubenswrapper[4735]: I1122 08:41:27.271579 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:41:27 crc kubenswrapper[4735]: I1122 08:41:27.275956 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hd8r" event={"ID":"48abc3dc-cfb2-4f33-bf53-17d1a069bc46","Type":"ContainerDied","Data":"24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26"} Nov 22 08:41:27 crc kubenswrapper[4735]: I1122 08:41:27.276000 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hd8r" event={"ID":"48abc3dc-cfb2-4f33-bf53-17d1a069bc46","Type":"ContainerStarted","Data":"441efa4100f9567d29efeb875a1bca2e0ed71f1fe0f87114cb482717a448ce9c"} Nov 22 08:41:28 crc kubenswrapper[4735]: I1122 08:41:28.282994 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hd8r" event={"ID":"48abc3dc-cfb2-4f33-bf53-17d1a069bc46","Type":"ContainerStarted","Data":"a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b"} Nov 22 08:41:33 crc kubenswrapper[4735]: I1122 08:41:33.333107 4735 generic.go:334] "Generic (PLEG): container finished" podID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerID="a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b" exitCode=0 Nov 22 08:41:33 crc kubenswrapper[4735]: I1122 08:41:33.333188 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hd8r" event={"ID":"48abc3dc-cfb2-4f33-bf53-17d1a069bc46","Type":"ContainerDied","Data":"a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b"} Nov 22 08:41:34 crc kubenswrapper[4735]: I1122 08:41:34.348245 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hd8r" event={"ID":"48abc3dc-cfb2-4f33-bf53-17d1a069bc46","Type":"ContainerStarted","Data":"2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6"} Nov 22 08:41:34 crc kubenswrapper[4735]: I1122 08:41:34.351146 4735 generic.go:334] "Generic (PLEG): container finished" podID="311958ee-5838-4326-9b40-7b7228bd6393" containerID="d0ddda72651d8d2964a566ec27b26c79eabf2591dad5ea4a74fe71dc813dd954" exitCode=0 Nov 22 08:41:34 crc kubenswrapper[4735]: I1122 08:41:34.351208 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" event={"ID":"311958ee-5838-4326-9b40-7b7228bd6393","Type":"ContainerDied","Data":"d0ddda72651d8d2964a566ec27b26c79eabf2591dad5ea4a74fe71dc813dd954"} Nov 22 08:41:34 crc kubenswrapper[4735]: I1122 08:41:34.372362 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8hd8r" podStartSLOduration=2.88068618 podStartE2EDuration="9.37234127s" podCreationTimestamp="2025-11-22 08:41:25 +0000 UTC" firstStartedPulling="2025-11-22 08:41:27.271272238 +0000 UTC m=+2308.875610853" lastFinishedPulling="2025-11-22 08:41:33.762927338 +0000 UTC m=+2315.367265943" observedRunningTime="2025-11-22 08:41:34.369826835 +0000 UTC m=+2315.974165460" watchObservedRunningTime="2025-11-22 08:41:34.37234127 +0000 UTC m=+2315.976679875" Nov 22 08:41:35 crc kubenswrapper[4735]: I1122 08:41:35.833431 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:35 crc kubenswrapper[4735]: I1122 08:41:35.833774 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:35 crc kubenswrapper[4735]: I1122 08:41:35.941063 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:41:35 crc kubenswrapper[4735]: I1122 08:41:35.958003 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/311958ee-5838-4326-9b40-7b7228bd6393-ovncontroller-config-0\") pod \"311958ee-5838-4326-9b40-7b7228bd6393\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " Nov 22 08:41:35 crc kubenswrapper[4735]: I1122 08:41:35.958135 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-inventory\") pod \"311958ee-5838-4326-9b40-7b7228bd6393\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " Nov 22 08:41:35 crc kubenswrapper[4735]: I1122 08:41:35.958291 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ovn-combined-ca-bundle\") pod \"311958ee-5838-4326-9b40-7b7228bd6393\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " Nov 22 08:41:35 crc kubenswrapper[4735]: I1122 08:41:35.958402 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ssh-key\") pod \"311958ee-5838-4326-9b40-7b7228bd6393\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " Nov 22 08:41:35 crc kubenswrapper[4735]: I1122 08:41:35.968973 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "311958ee-5838-4326-9b40-7b7228bd6393" (UID: "311958ee-5838-4326-9b40-7b7228bd6393"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.014916 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-inventory" (OuterVolumeSpecName: "inventory") pod "311958ee-5838-4326-9b40-7b7228bd6393" (UID: "311958ee-5838-4326-9b40-7b7228bd6393"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.041951 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/311958ee-5838-4326-9b40-7b7228bd6393-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "311958ee-5838-4326-9b40-7b7228bd6393" (UID: "311958ee-5838-4326-9b40-7b7228bd6393"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.043288 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "311958ee-5838-4326-9b40-7b7228bd6393" (UID: "311958ee-5838-4326-9b40-7b7228bd6393"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.062062 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4txwd\" (UniqueName: \"kubernetes.io/projected/311958ee-5838-4326-9b40-7b7228bd6393-kube-api-access-4txwd\") pod \"311958ee-5838-4326-9b40-7b7228bd6393\" (UID: \"311958ee-5838-4326-9b40-7b7228bd6393\") " Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.062776 4735 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/311958ee-5838-4326-9b40-7b7228bd6393-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.062795 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.062804 4735 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.062813 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311958ee-5838-4326-9b40-7b7228bd6393-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.066167 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/311958ee-5838-4326-9b40-7b7228bd6393-kube-api-access-4txwd" (OuterVolumeSpecName: "kube-api-access-4txwd") pod "311958ee-5838-4326-9b40-7b7228bd6393" (UID: "311958ee-5838-4326-9b40-7b7228bd6393"). InnerVolumeSpecName "kube-api-access-4txwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.165590 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4txwd\" (UniqueName: \"kubernetes.io/projected/311958ee-5838-4326-9b40-7b7228bd6393-kube-api-access-4txwd\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.378209 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" event={"ID":"311958ee-5838-4326-9b40-7b7228bd6393","Type":"ContainerDied","Data":"006cbdb1c894eaf7638bf3276af1e55d505d1328fb4350c5d85ffbcddd2fcd6c"} Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.378261 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="006cbdb1c894eaf7638bf3276af1e55d505d1328fb4350c5d85ffbcddd2fcd6c" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.378320 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lgnq6" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.519321 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2"] Nov 22 08:41:36 crc kubenswrapper[4735]: E1122 08:41:36.519935 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="311958ee-5838-4326-9b40-7b7228bd6393" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.520217 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="311958ee-5838-4326-9b40-7b7228bd6393" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.520534 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="311958ee-5838-4326-9b40-7b7228bd6393" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.521363 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.523943 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.523953 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.523960 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.524620 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.525269 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.525517 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.533139 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2"] Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.709979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.710095 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.710124 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.710186 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.710220 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2rmw\" (UniqueName: \"kubernetes.io/projected/06f78f85-aa4b-41bc-8b12-3fefeca970ee-kube-api-access-k2rmw\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.710280 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.812114 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.812225 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.812341 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.812378 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.812484 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.812522 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2rmw\" (UniqueName: \"kubernetes.io/projected/06f78f85-aa4b-41bc-8b12-3fefeca970ee-kube-api-access-k2rmw\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.817829 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.818206 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.818321 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.819742 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.828240 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.835285 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2rmw\" (UniqueName: \"kubernetes.io/projected/06f78f85-aa4b-41bc-8b12-3fefeca970ee-kube-api-access-k2rmw\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.837494 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:41:36 crc kubenswrapper[4735]: I1122 08:41:36.887134 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8hd8r" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="registry-server" probeResult="failure" output=< Nov 22 08:41:36 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:41:36 crc kubenswrapper[4735]: > Nov 22 08:41:37 crc kubenswrapper[4735]: I1122 08:41:37.391447 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2"] Nov 22 08:41:37 crc kubenswrapper[4735]: W1122 08:41:37.403176 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06f78f85_aa4b_41bc_8b12_3fefeca970ee.slice/crio-d6024d3fbe4c6fc223e8bb859681a53a4c598cf1e68770dea8781600ff8eb644 WatchSource:0}: Error finding container d6024d3fbe4c6fc223e8bb859681a53a4c598cf1e68770dea8781600ff8eb644: Status 404 returned error can't find the container with id d6024d3fbe4c6fc223e8bb859681a53a4c598cf1e68770dea8781600ff8eb644 Nov 22 08:41:38 crc kubenswrapper[4735]: I1122 08:41:38.402317 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" event={"ID":"06f78f85-aa4b-41bc-8b12-3fefeca970ee","Type":"ContainerStarted","Data":"e18cb5913f091a481af1e3a84294378f1e406a7205a1dad896c3a59fa21c8647"} Nov 22 08:41:38 crc kubenswrapper[4735]: I1122 08:41:38.402621 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" event={"ID":"06f78f85-aa4b-41bc-8b12-3fefeca970ee","Type":"ContainerStarted","Data":"d6024d3fbe4c6fc223e8bb859681a53a4c598cf1e68770dea8781600ff8eb644"} Nov 22 08:41:38 crc kubenswrapper[4735]: I1122 08:41:38.425142 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" podStartSLOduration=1.942992955 podStartE2EDuration="2.425124698s" podCreationTimestamp="2025-11-22 08:41:36 +0000 UTC" firstStartedPulling="2025-11-22 08:41:37.406927428 +0000 UTC m=+2319.011266033" lastFinishedPulling="2025-11-22 08:41:37.889059171 +0000 UTC m=+2319.493397776" observedRunningTime="2025-11-22 08:41:38.419868421 +0000 UTC m=+2320.024207026" watchObservedRunningTime="2025-11-22 08:41:38.425124698 +0000 UTC m=+2320.029463303" Nov 22 08:41:39 crc kubenswrapper[4735]: I1122 08:41:39.263201 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:41:39 crc kubenswrapper[4735]: E1122 08:41:39.263722 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:41:45 crc kubenswrapper[4735]: I1122 08:41:45.900281 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:45 crc kubenswrapper[4735]: I1122 08:41:45.951379 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:46 crc kubenswrapper[4735]: I1122 08:41:46.153990 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8hd8r"] Nov 22 08:41:47 crc kubenswrapper[4735]: I1122 08:41:47.492744 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8hd8r" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="registry-server" containerID="cri-o://2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6" gracePeriod=2 Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.052281 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.084924 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-catalog-content\") pod \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.084977 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx6sb\" (UniqueName: \"kubernetes.io/projected/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-kube-api-access-sx6sb\") pod \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.093789 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-kube-api-access-sx6sb" (OuterVolumeSpecName: "kube-api-access-sx6sb") pod "48abc3dc-cfb2-4f33-bf53-17d1a069bc46" (UID: "48abc3dc-cfb2-4f33-bf53-17d1a069bc46"). InnerVolumeSpecName "kube-api-access-sx6sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.186574 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-utilities\") pod \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\" (UID: \"48abc3dc-cfb2-4f33-bf53-17d1a069bc46\") " Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.187168 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx6sb\" (UniqueName: \"kubernetes.io/projected/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-kube-api-access-sx6sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.187336 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-utilities" (OuterVolumeSpecName: "utilities") pod "48abc3dc-cfb2-4f33-bf53-17d1a069bc46" (UID: "48abc3dc-cfb2-4f33-bf53-17d1a069bc46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.208810 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48abc3dc-cfb2-4f33-bf53-17d1a069bc46" (UID: "48abc3dc-cfb2-4f33-bf53-17d1a069bc46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.291392 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.291429 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48abc3dc-cfb2-4f33-bf53-17d1a069bc46-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.505025 4735 generic.go:334] "Generic (PLEG): container finished" podID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerID="2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6" exitCode=0 Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.505100 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hd8r" event={"ID":"48abc3dc-cfb2-4f33-bf53-17d1a069bc46","Type":"ContainerDied","Data":"2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6"} Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.505155 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hd8r" event={"ID":"48abc3dc-cfb2-4f33-bf53-17d1a069bc46","Type":"ContainerDied","Data":"441efa4100f9567d29efeb875a1bca2e0ed71f1fe0f87114cb482717a448ce9c"} Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.505183 4735 scope.go:117] "RemoveContainer" containerID="2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.505441 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hd8r" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.539564 4735 scope.go:117] "RemoveContainer" containerID="a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.544056 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8hd8r"] Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.565331 4735 scope.go:117] "RemoveContainer" containerID="24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.571208 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8hd8r"] Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.627360 4735 scope.go:117] "RemoveContainer" containerID="2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6" Nov 22 08:41:48 crc kubenswrapper[4735]: E1122 08:41:48.628156 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6\": container with ID starting with 2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6 not found: ID does not exist" containerID="2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.628218 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6"} err="failed to get container status \"2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6\": rpc error: code = NotFound desc = could not find container \"2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6\": container with ID starting with 2e066d09677f88969799498cb04ee669e3e94046e3465cb8f1dc1ddc9c0f2cf6 not found: ID does not exist" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.628258 4735 scope.go:117] "RemoveContainer" containerID="a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b" Nov 22 08:41:48 crc kubenswrapper[4735]: E1122 08:41:48.628999 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b\": container with ID starting with a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b not found: ID does not exist" containerID="a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.629092 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b"} err="failed to get container status \"a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b\": rpc error: code = NotFound desc = could not find container \"a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b\": container with ID starting with a06a5c06e0aec17b42db379f475ccd2431d60313e14666393a086170eee5176b not found: ID does not exist" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.629121 4735 scope.go:117] "RemoveContainer" containerID="24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26" Nov 22 08:41:48 crc kubenswrapper[4735]: E1122 08:41:48.630837 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26\": container with ID starting with 24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26 not found: ID does not exist" containerID="24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26" Nov 22 08:41:48 crc kubenswrapper[4735]: I1122 08:41:48.630880 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26"} err="failed to get container status \"24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26\": rpc error: code = NotFound desc = could not find container \"24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26\": container with ID starting with 24484dad2075460a36ab39f175b065ba7465e3c6308f2e4437eae33cf4e8df26 not found: ID does not exist" Nov 22 08:41:49 crc kubenswrapper[4735]: I1122 08:41:49.283689 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" path="/var/lib/kubelet/pods/48abc3dc-cfb2-4f33-bf53-17d1a069bc46/volumes" Nov 22 08:41:53 crc kubenswrapper[4735]: I1122 08:41:53.263680 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:41:53 crc kubenswrapper[4735]: E1122 08:41:53.264509 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:42:07 crc kubenswrapper[4735]: I1122 08:42:07.264665 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:42:07 crc kubenswrapper[4735]: E1122 08:42:07.266233 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.123192 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vcbwk"] Nov 22 08:42:08 crc kubenswrapper[4735]: E1122 08:42:08.124418 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="extract-utilities" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.124550 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="extract-utilities" Nov 22 08:42:08 crc kubenswrapper[4735]: E1122 08:42:08.124717 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="extract-content" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.124792 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="extract-content" Nov 22 08:42:08 crc kubenswrapper[4735]: E1122 08:42:08.124871 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="registry-server" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.124950 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="registry-server" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.125337 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="48abc3dc-cfb2-4f33-bf53-17d1a069bc46" containerName="registry-server" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.129698 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.136481 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vcbwk"] Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.178781 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-catalog-content\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.179322 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-utilities\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.179507 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt6xg\" (UniqueName: \"kubernetes.io/projected/8c6ea0d5-1638-4084-bc82-f41d460e3daa-kube-api-access-vt6xg\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.281935 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-utilities\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.282018 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt6xg\" (UniqueName: \"kubernetes.io/projected/8c6ea0d5-1638-4084-bc82-f41d460e3daa-kube-api-access-vt6xg\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.282083 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-catalog-content\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.282475 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-utilities\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.282743 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-catalog-content\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.301507 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt6xg\" (UniqueName: \"kubernetes.io/projected/8c6ea0d5-1638-4084-bc82-f41d460e3daa-kube-api-access-vt6xg\") pod \"community-operators-vcbwk\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:08 crc kubenswrapper[4735]: I1122 08:42:08.455607 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:09 crc kubenswrapper[4735]: I1122 08:42:09.570118 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vcbwk"] Nov 22 08:42:09 crc kubenswrapper[4735]: I1122 08:42:09.796977 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcbwk" event={"ID":"8c6ea0d5-1638-4084-bc82-f41d460e3daa","Type":"ContainerStarted","Data":"7caa5ed6b35732cb5690bfdbd78c69af99e4265606aa67d1b11f0432cbd63823"} Nov 22 08:42:10 crc kubenswrapper[4735]: I1122 08:42:10.811088 4735 generic.go:334] "Generic (PLEG): container finished" podID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerID="b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d" exitCode=0 Nov 22 08:42:10 crc kubenswrapper[4735]: I1122 08:42:10.811275 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcbwk" event={"ID":"8c6ea0d5-1638-4084-bc82-f41d460e3daa","Type":"ContainerDied","Data":"b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d"} Nov 22 08:42:11 crc kubenswrapper[4735]: I1122 08:42:11.822742 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcbwk" event={"ID":"8c6ea0d5-1638-4084-bc82-f41d460e3daa","Type":"ContainerStarted","Data":"2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432"} Nov 22 08:42:13 crc kubenswrapper[4735]: I1122 08:42:13.857414 4735 generic.go:334] "Generic (PLEG): container finished" podID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerID="2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432" exitCode=0 Nov 22 08:42:13 crc kubenswrapper[4735]: I1122 08:42:13.857519 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcbwk" event={"ID":"8c6ea0d5-1638-4084-bc82-f41d460e3daa","Type":"ContainerDied","Data":"2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432"} Nov 22 08:42:14 crc kubenswrapper[4735]: I1122 08:42:14.872647 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcbwk" event={"ID":"8c6ea0d5-1638-4084-bc82-f41d460e3daa","Type":"ContainerStarted","Data":"df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25"} Nov 22 08:42:14 crc kubenswrapper[4735]: I1122 08:42:14.894001 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vcbwk" podStartSLOduration=3.398457174 podStartE2EDuration="6.893980384s" podCreationTimestamp="2025-11-22 08:42:08 +0000 UTC" firstStartedPulling="2025-11-22 08:42:10.81399186 +0000 UTC m=+2352.418330455" lastFinishedPulling="2025-11-22 08:42:14.30951505 +0000 UTC m=+2355.913853665" observedRunningTime="2025-11-22 08:42:14.891003157 +0000 UTC m=+2356.495341772" watchObservedRunningTime="2025-11-22 08:42:14.893980384 +0000 UTC m=+2356.498318999" Nov 22 08:42:18 crc kubenswrapper[4735]: I1122 08:42:18.457531 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:18 crc kubenswrapper[4735]: I1122 08:42:18.457948 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:18 crc kubenswrapper[4735]: I1122 08:42:18.547474 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:21 crc kubenswrapper[4735]: I1122 08:42:21.262773 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:42:21 crc kubenswrapper[4735]: E1122 08:42:21.263335 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:42:28 crc kubenswrapper[4735]: I1122 08:42:28.507420 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:28 crc kubenswrapper[4735]: I1122 08:42:28.555503 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vcbwk"] Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.059529 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vcbwk" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerName="registry-server" containerID="cri-o://df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25" gracePeriod=2 Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.702279 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.766709 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-catalog-content\") pod \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.766784 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt6xg\" (UniqueName: \"kubernetes.io/projected/8c6ea0d5-1638-4084-bc82-f41d460e3daa-kube-api-access-vt6xg\") pod \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.766974 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-utilities\") pod \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\" (UID: \"8c6ea0d5-1638-4084-bc82-f41d460e3daa\") " Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.767825 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-utilities" (OuterVolumeSpecName: "utilities") pod "8c6ea0d5-1638-4084-bc82-f41d460e3daa" (UID: "8c6ea0d5-1638-4084-bc82-f41d460e3daa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.779758 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c6ea0d5-1638-4084-bc82-f41d460e3daa-kube-api-access-vt6xg" (OuterVolumeSpecName: "kube-api-access-vt6xg") pod "8c6ea0d5-1638-4084-bc82-f41d460e3daa" (UID: "8c6ea0d5-1638-4084-bc82-f41d460e3daa"). InnerVolumeSpecName "kube-api-access-vt6xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.824356 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c6ea0d5-1638-4084-bc82-f41d460e3daa" (UID: "8c6ea0d5-1638-4084-bc82-f41d460e3daa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.869523 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.869558 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6ea0d5-1638-4084-bc82-f41d460e3daa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:29 crc kubenswrapper[4735]: I1122 08:42:29.869571 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt6xg\" (UniqueName: \"kubernetes.io/projected/8c6ea0d5-1638-4084-bc82-f41d460e3daa-kube-api-access-vt6xg\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.071983 4735 generic.go:334] "Generic (PLEG): container finished" podID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerID="df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25" exitCode=0 Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.072030 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcbwk" event={"ID":"8c6ea0d5-1638-4084-bc82-f41d460e3daa","Type":"ContainerDied","Data":"df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25"} Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.072043 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vcbwk" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.072058 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vcbwk" event={"ID":"8c6ea0d5-1638-4084-bc82-f41d460e3daa","Type":"ContainerDied","Data":"7caa5ed6b35732cb5690bfdbd78c69af99e4265606aa67d1b11f0432cbd63823"} Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.072082 4735 scope.go:117] "RemoveContainer" containerID="df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.106499 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vcbwk"] Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.106903 4735 scope.go:117] "RemoveContainer" containerID="2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.118118 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vcbwk"] Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.147707 4735 scope.go:117] "RemoveContainer" containerID="b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.187826 4735 scope.go:117] "RemoveContainer" containerID="df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25" Nov 22 08:42:30 crc kubenswrapper[4735]: E1122 08:42:30.188147 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25\": container with ID starting with df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25 not found: ID does not exist" containerID="df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.188172 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25"} err="failed to get container status \"df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25\": rpc error: code = NotFound desc = could not find container \"df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25\": container with ID starting with df1ab0bf17cd7a0988a1c7ab803393fdb5958c711474366ed512a1b83bb7ba25 not found: ID does not exist" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.188192 4735 scope.go:117] "RemoveContainer" containerID="2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432" Nov 22 08:42:30 crc kubenswrapper[4735]: E1122 08:42:30.188644 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432\": container with ID starting with 2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432 not found: ID does not exist" containerID="2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.188710 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432"} err="failed to get container status \"2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432\": rpc error: code = NotFound desc = could not find container \"2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432\": container with ID starting with 2f1e7893ff295d8e207c3237cc2e18e468eea14b20ec294f13661bc5603b9432 not found: ID does not exist" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.188751 4735 scope.go:117] "RemoveContainer" containerID="b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d" Nov 22 08:42:30 crc kubenswrapper[4735]: E1122 08:42:30.189427 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d\": container with ID starting with b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d not found: ID does not exist" containerID="b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d" Nov 22 08:42:30 crc kubenswrapper[4735]: I1122 08:42:30.189496 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d"} err="failed to get container status \"b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d\": rpc error: code = NotFound desc = could not find container \"b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d\": container with ID starting with b76b5178650fae0b431218fa1781ead20d5111eaadb4002d6e6440184f39073d not found: ID does not exist" Nov 22 08:42:31 crc kubenswrapper[4735]: I1122 08:42:31.282175 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" path="/var/lib/kubelet/pods/8c6ea0d5-1638-4084-bc82-f41d460e3daa/volumes" Nov 22 08:42:32 crc kubenswrapper[4735]: I1122 08:42:32.100089 4735 generic.go:334] "Generic (PLEG): container finished" podID="06f78f85-aa4b-41bc-8b12-3fefeca970ee" containerID="e18cb5913f091a481af1e3a84294378f1e406a7205a1dad896c3a59fa21c8647" exitCode=0 Nov 22 08:42:32 crc kubenswrapper[4735]: I1122 08:42:32.100174 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" event={"ID":"06f78f85-aa4b-41bc-8b12-3fefeca970ee","Type":"ContainerDied","Data":"e18cb5913f091a481af1e3a84294378f1e406a7205a1dad896c3a59fa21c8647"} Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.604434 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.768604 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-inventory\") pod \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.768761 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-metadata-combined-ca-bundle\") pod \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.768931 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-ovn-metadata-agent-neutron-config-0\") pod \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.768981 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2rmw\" (UniqueName: \"kubernetes.io/projected/06f78f85-aa4b-41bc-8b12-3fefeca970ee-kube-api-access-k2rmw\") pod \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.769140 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-nova-metadata-neutron-config-0\") pod \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.769178 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-ssh-key\") pod \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\" (UID: \"06f78f85-aa4b-41bc-8b12-3fefeca970ee\") " Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.773948 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f78f85-aa4b-41bc-8b12-3fefeca970ee-kube-api-access-k2rmw" (OuterVolumeSpecName: "kube-api-access-k2rmw") pod "06f78f85-aa4b-41bc-8b12-3fefeca970ee" (UID: "06f78f85-aa4b-41bc-8b12-3fefeca970ee"). InnerVolumeSpecName "kube-api-access-k2rmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.775630 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "06f78f85-aa4b-41bc-8b12-3fefeca970ee" (UID: "06f78f85-aa4b-41bc-8b12-3fefeca970ee"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.804800 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "06f78f85-aa4b-41bc-8b12-3fefeca970ee" (UID: "06f78f85-aa4b-41bc-8b12-3fefeca970ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.807640 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-inventory" (OuterVolumeSpecName: "inventory") pod "06f78f85-aa4b-41bc-8b12-3fefeca970ee" (UID: "06f78f85-aa4b-41bc-8b12-3fefeca970ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.811843 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "06f78f85-aa4b-41bc-8b12-3fefeca970ee" (UID: "06f78f85-aa4b-41bc-8b12-3fefeca970ee"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.817793 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "06f78f85-aa4b-41bc-8b12-3fefeca970ee" (UID: "06f78f85-aa4b-41bc-8b12-3fefeca970ee"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.872416 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2rmw\" (UniqueName: \"kubernetes.io/projected/06f78f85-aa4b-41bc-8b12-3fefeca970ee-kube-api-access-k2rmw\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.872493 4735 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.872513 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.872527 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.872539 4735 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:33 crc kubenswrapper[4735]: I1122 08:42:33.872557 4735 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06f78f85-aa4b-41bc-8b12-3fefeca970ee-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.123650 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" event={"ID":"06f78f85-aa4b-41bc-8b12-3fefeca970ee","Type":"ContainerDied","Data":"d6024d3fbe4c6fc223e8bb859681a53a4c598cf1e68770dea8781600ff8eb644"} Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.123938 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6024d3fbe4c6fc223e8bb859681a53a4c598cf1e68770dea8781600ff8eb644" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.123734 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.287641 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w"] Nov 22 08:42:34 crc kubenswrapper[4735]: E1122 08:42:34.288698 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f78f85-aa4b-41bc-8b12-3fefeca970ee" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.288722 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f78f85-aa4b-41bc-8b12-3fefeca970ee" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 08:42:34 crc kubenswrapper[4735]: E1122 08:42:34.288766 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerName="extract-utilities" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.288776 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerName="extract-utilities" Nov 22 08:42:34 crc kubenswrapper[4735]: E1122 08:42:34.288788 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerName="registry-server" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.288796 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerName="registry-server" Nov 22 08:42:34 crc kubenswrapper[4735]: E1122 08:42:34.288818 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerName="extract-content" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.288826 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerName="extract-content" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.289106 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c6ea0d5-1638-4084-bc82-f41d460e3daa" containerName="registry-server" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.289143 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f78f85-aa4b-41bc-8b12-3fefeca970ee" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.290086 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.292252 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.292674 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.292682 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.292719 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.297348 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.316075 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w"] Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.383731 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.383781 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkksh\" (UniqueName: \"kubernetes.io/projected/341a140a-5530-434a-b245-e07c4a7068fc-kube-api-access-kkksh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.383862 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.384022 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.384362 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.486263 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.486389 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.486429 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.486454 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkksh\" (UniqueName: \"kubernetes.io/projected/341a140a-5530-434a-b245-e07c4a7068fc-kube-api-access-kkksh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.486510 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.490221 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.490375 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.490860 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.498477 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.508638 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkksh\" (UniqueName: \"kubernetes.io/projected/341a140a-5530-434a-b245-e07c4a7068fc-kube-api-access-kkksh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:34 crc kubenswrapper[4735]: I1122 08:42:34.614873 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:42:35 crc kubenswrapper[4735]: W1122 08:42:35.230266 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod341a140a_5530_434a_b245_e07c4a7068fc.slice/crio-d36a970f4d2c11a182809b88e435821c79a7e2d22bad8c232afaac5ddeaf81ad WatchSource:0}: Error finding container d36a970f4d2c11a182809b88e435821c79a7e2d22bad8c232afaac5ddeaf81ad: Status 404 returned error can't find the container with id d36a970f4d2c11a182809b88e435821c79a7e2d22bad8c232afaac5ddeaf81ad Nov 22 08:42:35 crc kubenswrapper[4735]: I1122 08:42:35.238436 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w"] Nov 22 08:42:35 crc kubenswrapper[4735]: I1122 08:42:35.265122 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:42:35 crc kubenswrapper[4735]: E1122 08:42:35.265449 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:42:36 crc kubenswrapper[4735]: I1122 08:42:36.151588 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" event={"ID":"341a140a-5530-434a-b245-e07c4a7068fc","Type":"ContainerStarted","Data":"1fbe14fa520c558983416fcf1632944176fb870676d17c87a6c325ae811f25f9"} Nov 22 08:42:36 crc kubenswrapper[4735]: I1122 08:42:36.152007 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" event={"ID":"341a140a-5530-434a-b245-e07c4a7068fc","Type":"ContainerStarted","Data":"d36a970f4d2c11a182809b88e435821c79a7e2d22bad8c232afaac5ddeaf81ad"} Nov 22 08:42:36 crc kubenswrapper[4735]: I1122 08:42:36.185989 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" podStartSLOduration=1.652275905 podStartE2EDuration="2.185965198s" podCreationTimestamp="2025-11-22 08:42:34 +0000 UTC" firstStartedPulling="2025-11-22 08:42:35.23474737 +0000 UTC m=+2376.839085985" lastFinishedPulling="2025-11-22 08:42:35.768436643 +0000 UTC m=+2377.372775278" observedRunningTime="2025-11-22 08:42:36.177126498 +0000 UTC m=+2377.781465113" watchObservedRunningTime="2025-11-22 08:42:36.185965198 +0000 UTC m=+2377.790303813" Nov 22 08:42:50 crc kubenswrapper[4735]: I1122 08:42:50.264152 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:42:50 crc kubenswrapper[4735]: E1122 08:42:50.265139 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:43:04 crc kubenswrapper[4735]: I1122 08:43:04.265869 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:43:04 crc kubenswrapper[4735]: E1122 08:43:04.267032 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:43:15 crc kubenswrapper[4735]: I1122 08:43:15.264726 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:43:15 crc kubenswrapper[4735]: E1122 08:43:15.267111 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:43:30 crc kubenswrapper[4735]: I1122 08:43:30.266432 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:43:30 crc kubenswrapper[4735]: E1122 08:43:30.267253 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:43:42 crc kubenswrapper[4735]: I1122 08:43:42.264221 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:43:42 crc kubenswrapper[4735]: E1122 08:43:42.267659 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:43:55 crc kubenswrapper[4735]: I1122 08:43:55.264815 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:43:55 crc kubenswrapper[4735]: E1122 08:43:55.266437 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:44:06 crc kubenswrapper[4735]: I1122 08:44:06.263883 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:44:06 crc kubenswrapper[4735]: E1122 08:44:06.265105 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:44:19 crc kubenswrapper[4735]: I1122 08:44:19.270645 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:44:19 crc kubenswrapper[4735]: E1122 08:44:19.271634 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:44:33 crc kubenswrapper[4735]: I1122 08:44:33.264193 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:44:33 crc kubenswrapper[4735]: E1122 08:44:33.265065 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:44:44 crc kubenswrapper[4735]: I1122 08:44:44.264015 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:44:44 crc kubenswrapper[4735]: E1122 08:44:44.265985 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:44:57 crc kubenswrapper[4735]: I1122 08:44:57.263599 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:44:57 crc kubenswrapper[4735]: E1122 08:44:57.264321 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.150941 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d"] Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.156011 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.161089 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.161162 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.183096 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d"] Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.321639 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-config-volume\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.322216 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-secret-volume\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.322382 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbg2d\" (UniqueName: \"kubernetes.io/projected/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-kube-api-access-pbg2d\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.425409 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-secret-volume\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.425567 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbg2d\" (UniqueName: \"kubernetes.io/projected/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-kube-api-access-pbg2d\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.425867 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-config-volume\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.426977 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-config-volume\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.432369 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-secret-volume\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.444856 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbg2d\" (UniqueName: \"kubernetes.io/projected/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-kube-api-access-pbg2d\") pod \"collect-profiles-29396685-rng9d\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:00 crc kubenswrapper[4735]: I1122 08:45:00.484590 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:01 crc kubenswrapper[4735]: I1122 08:45:01.004164 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d"] Nov 22 08:45:01 crc kubenswrapper[4735]: I1122 08:45:01.911830 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" event={"ID":"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8","Type":"ContainerStarted","Data":"e71a621fbeea5d717029f1490c681f4e9cadffeefe63d3dba96a90050fccf2cb"} Nov 22 08:45:01 crc kubenswrapper[4735]: I1122 08:45:01.912156 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" event={"ID":"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8","Type":"ContainerStarted","Data":"a041795f0db38437d2ba1bb3c4d81d41e250b9f8dcd1cb38af9a4c13c3337b68"} Nov 22 08:45:01 crc kubenswrapper[4735]: I1122 08:45:01.932819 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" podStartSLOduration=1.932780889 podStartE2EDuration="1.932780889s" podCreationTimestamp="2025-11-22 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:45:01.925689366 +0000 UTC m=+2523.530027991" watchObservedRunningTime="2025-11-22 08:45:01.932780889 +0000 UTC m=+2523.537119524" Nov 22 08:45:02 crc kubenswrapper[4735]: I1122 08:45:02.924852 4735 generic.go:334] "Generic (PLEG): container finished" podID="45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8" containerID="e71a621fbeea5d717029f1490c681f4e9cadffeefe63d3dba96a90050fccf2cb" exitCode=0 Nov 22 08:45:02 crc kubenswrapper[4735]: I1122 08:45:02.924925 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" event={"ID":"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8","Type":"ContainerDied","Data":"e71a621fbeea5d717029f1490c681f4e9cadffeefe63d3dba96a90050fccf2cb"} Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.368054 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.470713 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-config-volume\") pod \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.471019 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbg2d\" (UniqueName: \"kubernetes.io/projected/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-kube-api-access-pbg2d\") pod \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.471660 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-config-volume" (OuterVolumeSpecName: "config-volume") pod "45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8" (UID: "45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.471951 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-secret-volume\") pod \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\" (UID: \"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8\") " Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.473234 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.477414 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-kube-api-access-pbg2d" (OuterVolumeSpecName: "kube-api-access-pbg2d") pod "45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8" (UID: "45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8"). InnerVolumeSpecName "kube-api-access-pbg2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.482775 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8" (UID: "45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.575169 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbg2d\" (UniqueName: \"kubernetes.io/projected/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-kube-api-access-pbg2d\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.575209 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.954011 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" event={"ID":"45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8","Type":"ContainerDied","Data":"a041795f0db38437d2ba1bb3c4d81d41e250b9f8dcd1cb38af9a4c13c3337b68"} Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.954072 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a041795f0db38437d2ba1bb3c4d81d41e250b9f8dcd1cb38af9a4c13c3337b68" Nov 22 08:45:04 crc kubenswrapper[4735]: I1122 08:45:04.954142 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d" Nov 22 08:45:05 crc kubenswrapper[4735]: I1122 08:45:05.012332 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q"] Nov 22 08:45:05 crc kubenswrapper[4735]: I1122 08:45:05.023598 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-xdk4q"] Nov 22 08:45:05 crc kubenswrapper[4735]: I1122 08:45:05.289416 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31a557ae-6002-49fd-a9af-b3ea23dfd952" path="/var/lib/kubelet/pods/31a557ae-6002-49fd-a9af-b3ea23dfd952/volumes" Nov 22 08:45:08 crc kubenswrapper[4735]: I1122 08:45:08.907605 4735 scope.go:117] "RemoveContainer" containerID="edf114b0a6ecb80547997ca80f1020152ca9adb4047e825a975d92c607a1f28c" Nov 22 08:45:12 crc kubenswrapper[4735]: I1122 08:45:12.264552 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:45:12 crc kubenswrapper[4735]: E1122 08:45:12.266204 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:45:26 crc kubenswrapper[4735]: I1122 08:45:26.263444 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:45:27 crc kubenswrapper[4735]: I1122 08:45:27.214479 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"fd09bda8cc7301147359675a9e8663a853932b4858c0e3f7f11321068fb04ea8"} Nov 22 08:47:24 crc kubenswrapper[4735]: I1122 08:47:24.625503 4735 generic.go:334] "Generic (PLEG): container finished" podID="341a140a-5530-434a-b245-e07c4a7068fc" containerID="1fbe14fa520c558983416fcf1632944176fb870676d17c87a6c325ae811f25f9" exitCode=0 Nov 22 08:47:24 crc kubenswrapper[4735]: I1122 08:47:24.626125 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" event={"ID":"341a140a-5530-434a-b245-e07c4a7068fc","Type":"ContainerDied","Data":"1fbe14fa520c558983416fcf1632944176fb870676d17c87a6c325ae811f25f9"} Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.096477 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.169476 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-combined-ca-bundle\") pod \"341a140a-5530-434a-b245-e07c4a7068fc\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.169660 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkksh\" (UniqueName: \"kubernetes.io/projected/341a140a-5530-434a-b245-e07c4a7068fc-kube-api-access-kkksh\") pod \"341a140a-5530-434a-b245-e07c4a7068fc\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.169830 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-secret-0\") pod \"341a140a-5530-434a-b245-e07c4a7068fc\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.169859 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-ssh-key\") pod \"341a140a-5530-434a-b245-e07c4a7068fc\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.169929 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-inventory\") pod \"341a140a-5530-434a-b245-e07c4a7068fc\" (UID: \"341a140a-5530-434a-b245-e07c4a7068fc\") " Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.176953 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "341a140a-5530-434a-b245-e07c4a7068fc" (UID: "341a140a-5530-434a-b245-e07c4a7068fc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.185842 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341a140a-5530-434a-b245-e07c4a7068fc-kube-api-access-kkksh" (OuterVolumeSpecName: "kube-api-access-kkksh") pod "341a140a-5530-434a-b245-e07c4a7068fc" (UID: "341a140a-5530-434a-b245-e07c4a7068fc"). InnerVolumeSpecName "kube-api-access-kkksh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.208593 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-inventory" (OuterVolumeSpecName: "inventory") pod "341a140a-5530-434a-b245-e07c4a7068fc" (UID: "341a140a-5530-434a-b245-e07c4a7068fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.208814 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "341a140a-5530-434a-b245-e07c4a7068fc" (UID: "341a140a-5530-434a-b245-e07c4a7068fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.208963 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "341a140a-5530-434a-b245-e07c4a7068fc" (UID: "341a140a-5530-434a-b245-e07c4a7068fc"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.273033 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkksh\" (UniqueName: \"kubernetes.io/projected/341a140a-5530-434a-b245-e07c4a7068fc-kube-api-access-kkksh\") on node \"crc\" DevicePath \"\"" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.273068 4735 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.273078 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.273089 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.273099 4735 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341a140a-5530-434a-b245-e07c4a7068fc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.649094 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" event={"ID":"341a140a-5530-434a-b245-e07c4a7068fc","Type":"ContainerDied","Data":"d36a970f4d2c11a182809b88e435821c79a7e2d22bad8c232afaac5ddeaf81ad"} Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.649137 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d36a970f4d2c11a182809b88e435821c79a7e2d22bad8c232afaac5ddeaf81ad" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.649213 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.751790 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns"] Nov 22 08:47:26 crc kubenswrapper[4735]: E1122 08:47:26.752753 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8" containerName="collect-profiles" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.752778 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8" containerName="collect-profiles" Nov 22 08:47:26 crc kubenswrapper[4735]: E1122 08:47:26.752819 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341a140a-5530-434a-b245-e07c4a7068fc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.752829 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="341a140a-5530-434a-b245-e07c4a7068fc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.753167 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="341a140a-5530-434a-b245-e07c4a7068fc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.753216 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8" containerName="collect-profiles" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.754278 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.757575 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.757596 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.757722 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.757722 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.757891 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.758128 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.763766 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns"] Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.771359 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.800993 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.801049 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.801090 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.801150 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.801183 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.801274 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.801339 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7f8k\" (UniqueName: \"kubernetes.io/projected/1867df3c-9d3b-41b0-9d6f-153536f35dd8-kube-api-access-z7f8k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.801570 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.801618 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.903443 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.903756 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.903947 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.904101 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7f8k\" (UniqueName: \"kubernetes.io/projected/1867df3c-9d3b-41b0-9d6f-153536f35dd8-kube-api-access-z7f8k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.904319 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.904438 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.904628 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.904752 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.904893 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.904769 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.907675 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.908129 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.908546 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.908572 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.908900 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.909915 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.910447 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:26 crc kubenswrapper[4735]: I1122 08:47:26.937367 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7f8k\" (UniqueName: \"kubernetes.io/projected/1867df3c-9d3b-41b0-9d6f-153536f35dd8-kube-api-access-z7f8k\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cf6ns\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:27 crc kubenswrapper[4735]: I1122 08:47:27.077339 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:47:27 crc kubenswrapper[4735]: I1122 08:47:27.724122 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:47:27 crc kubenswrapper[4735]: I1122 08:47:27.724800 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns"] Nov 22 08:47:28 crc kubenswrapper[4735]: I1122 08:47:28.670401 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" event={"ID":"1867df3c-9d3b-41b0-9d6f-153536f35dd8","Type":"ContainerStarted","Data":"74c9b0104def2d65c86780f0eee5343ff1c9768588b7f957d08be3c4bbbedb07"} Nov 22 08:47:28 crc kubenswrapper[4735]: I1122 08:47:28.670880 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" event={"ID":"1867df3c-9d3b-41b0-9d6f-153536f35dd8","Type":"ContainerStarted","Data":"0b926ad2b05ccd32b496005cd05c7373838b7b1399d970754808f7d69993019a"} Nov 22 08:47:46 crc kubenswrapper[4735]: I1122 08:47:46.131839 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:47:46 crc kubenswrapper[4735]: I1122 08:47:46.132435 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:48:16 crc kubenswrapper[4735]: I1122 08:48:16.131905 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:48:16 crc kubenswrapper[4735]: I1122 08:48:16.134578 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.131603 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.132700 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.132789 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.135456 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd09bda8cc7301147359675a9e8663a853932b4858c0e3f7f11321068fb04ea8"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.135642 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://fd09bda8cc7301147359675a9e8663a853932b4858c0e3f7f11321068fb04ea8" gracePeriod=600 Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.607660 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="fd09bda8cc7301147359675a9e8663a853932b4858c0e3f7f11321068fb04ea8" exitCode=0 Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.607723 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"fd09bda8cc7301147359675a9e8663a853932b4858c0e3f7f11321068fb04ea8"} Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.608033 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2"} Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.608062 4735 scope.go:117] "RemoveContainer" containerID="4ae6f8e3947168220cb177e15b799b49db00cad6a467ef4313a38963b78ec4a9" Nov 22 08:48:46 crc kubenswrapper[4735]: I1122 08:48:46.635288 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" podStartSLOduration=80.18662114 podStartE2EDuration="1m20.635268013s" podCreationTimestamp="2025-11-22 08:47:26 +0000 UTC" firstStartedPulling="2025-11-22 08:47:27.72386968 +0000 UTC m=+2669.328208285" lastFinishedPulling="2025-11-22 08:47:28.172516553 +0000 UTC m=+2669.776855158" observedRunningTime="2025-11-22 08:47:28.700404951 +0000 UTC m=+2670.304743566" watchObservedRunningTime="2025-11-22 08:48:46.635268013 +0000 UTC m=+2748.239606628" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.462941 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzsp"] Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.467228 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.491691 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzsp"] Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.595590 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-utilities\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.596003 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mskvt\" (UniqueName: \"kubernetes.io/projected/ef15b0b6-7e6e-4a09-93ad-64137cee3543-kube-api-access-mskvt\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.596139 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-catalog-content\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.698339 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-utilities\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.698793 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mskvt\" (UniqueName: \"kubernetes.io/projected/ef15b0b6-7e6e-4a09-93ad-64137cee3543-kube-api-access-mskvt\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.698916 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-catalog-content\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.699845 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-catalog-content\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.700115 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-utilities\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.722571 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mskvt\" (UniqueName: \"kubernetes.io/projected/ef15b0b6-7e6e-4a09-93ad-64137cee3543-kube-api-access-mskvt\") pod \"redhat-marketplace-xfzsp\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:07 crc kubenswrapper[4735]: I1122 08:50:07.802776 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:08 crc kubenswrapper[4735]: I1122 08:50:08.347882 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzsp"] Nov 22 08:50:08 crc kubenswrapper[4735]: I1122 08:50:08.541064 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzsp" event={"ID":"ef15b0b6-7e6e-4a09-93ad-64137cee3543","Type":"ContainerStarted","Data":"c65fd177e4c0b0a1998fbc112612b2791b3aea267a1e61909e600846f3755b78"} Nov 22 08:50:09 crc kubenswrapper[4735]: I1122 08:50:09.554430 4735 generic.go:334] "Generic (PLEG): container finished" podID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerID="1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156" exitCode=0 Nov 22 08:50:09 crc kubenswrapper[4735]: I1122 08:50:09.555218 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzsp" event={"ID":"ef15b0b6-7e6e-4a09-93ad-64137cee3543","Type":"ContainerDied","Data":"1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156"} Nov 22 08:50:10 crc kubenswrapper[4735]: I1122 08:50:10.569968 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzsp" event={"ID":"ef15b0b6-7e6e-4a09-93ad-64137cee3543","Type":"ContainerStarted","Data":"5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f"} Nov 22 08:50:11 crc kubenswrapper[4735]: I1122 08:50:11.582053 4735 generic.go:334] "Generic (PLEG): container finished" podID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerID="5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f" exitCode=0 Nov 22 08:50:11 crc kubenswrapper[4735]: I1122 08:50:11.582123 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzsp" event={"ID":"ef15b0b6-7e6e-4a09-93ad-64137cee3543","Type":"ContainerDied","Data":"5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f"} Nov 22 08:50:12 crc kubenswrapper[4735]: I1122 08:50:12.595454 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzsp" event={"ID":"ef15b0b6-7e6e-4a09-93ad-64137cee3543","Type":"ContainerStarted","Data":"731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721"} Nov 22 08:50:12 crc kubenswrapper[4735]: I1122 08:50:12.622333 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xfzsp" podStartSLOduration=2.88078004 podStartE2EDuration="5.622312062s" podCreationTimestamp="2025-11-22 08:50:07 +0000 UTC" firstStartedPulling="2025-11-22 08:50:09.557045267 +0000 UTC m=+2831.161383882" lastFinishedPulling="2025-11-22 08:50:12.298577299 +0000 UTC m=+2833.902915904" observedRunningTime="2025-11-22 08:50:12.613319304 +0000 UTC m=+2834.217657929" watchObservedRunningTime="2025-11-22 08:50:12.622312062 +0000 UTC m=+2834.226650657" Nov 22 08:50:17 crc kubenswrapper[4735]: I1122 08:50:17.803089 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:17 crc kubenswrapper[4735]: I1122 08:50:17.803393 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:17 crc kubenswrapper[4735]: I1122 08:50:17.871550 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:18 crc kubenswrapper[4735]: I1122 08:50:18.692354 4735 generic.go:334] "Generic (PLEG): container finished" podID="1867df3c-9d3b-41b0-9d6f-153536f35dd8" containerID="74c9b0104def2d65c86780f0eee5343ff1c9768588b7f957d08be3c4bbbedb07" exitCode=0 Nov 22 08:50:18 crc kubenswrapper[4735]: I1122 08:50:18.692421 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" event={"ID":"1867df3c-9d3b-41b0-9d6f-153536f35dd8","Type":"ContainerDied","Data":"74c9b0104def2d65c86780f0eee5343ff1c9768588b7f957d08be3c4bbbedb07"} Nov 22 08:50:18 crc kubenswrapper[4735]: I1122 08:50:18.745540 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:18 crc kubenswrapper[4735]: I1122 08:50:18.801056 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzsp"] Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.197215 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.332367 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-ssh-key\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.332721 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-1\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.332849 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-1\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.332877 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-inventory\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.332938 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7f8k\" (UniqueName: \"kubernetes.io/projected/1867df3c-9d3b-41b0-9d6f-153536f35dd8-kube-api-access-z7f8k\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.333151 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-combined-ca-bundle\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.333225 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-0\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.333291 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-extra-config-0\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.333335 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-0\") pod \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\" (UID: \"1867df3c-9d3b-41b0-9d6f-153536f35dd8\") " Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.340978 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1867df3c-9d3b-41b0-9d6f-153536f35dd8-kube-api-access-z7f8k" (OuterVolumeSpecName: "kube-api-access-z7f8k") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "kube-api-access-z7f8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.355389 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.367002 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.375578 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.376804 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.386735 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.405499 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.407134 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.409254 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-inventory" (OuterVolumeSpecName: "inventory") pod "1867df3c-9d3b-41b0-9d6f-153536f35dd8" (UID: "1867df3c-9d3b-41b0-9d6f-153536f35dd8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436027 4735 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436071 4735 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436087 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436099 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7f8k\" (UniqueName: \"kubernetes.io/projected/1867df3c-9d3b-41b0-9d6f-153536f35dd8-kube-api-access-z7f8k\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436112 4735 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436123 4735 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436135 4735 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436145 4735 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.436156 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1867df3c-9d3b-41b0-9d6f-153536f35dd8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.718003 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xfzsp" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerName="registry-server" containerID="cri-o://731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721" gracePeriod=2 Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.718497 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.720673 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cf6ns" event={"ID":"1867df3c-9d3b-41b0-9d6f-153536f35dd8","Type":"ContainerDied","Data":"0b926ad2b05ccd32b496005cd05c7373838b7b1399d970754808f7d69993019a"} Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.720717 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b926ad2b05ccd32b496005cd05c7373838b7b1399d970754808f7d69993019a" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.817752 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl"] Nov 22 08:50:20 crc kubenswrapper[4735]: E1122 08:50:20.818395 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1867df3c-9d3b-41b0-9d6f-153536f35dd8" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.818419 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="1867df3c-9d3b-41b0-9d6f-153536f35dd8" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.818670 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="1867df3c-9d3b-41b0-9d6f-153536f35dd8" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.819582 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.822146 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.822205 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.822213 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.822310 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.823622 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.827773 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl"] Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.948670 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.948751 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.948819 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.948930 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsbld\" (UniqueName: \"kubernetes.io/projected/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-kube-api-access-hsbld\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.948979 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.949415 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:20 crc kubenswrapper[4735]: I1122 08:50:20.949865 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.051762 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.051860 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.051914 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.051956 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.051999 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.052045 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsbld\" (UniqueName: \"kubernetes.io/projected/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-kube-api-access-hsbld\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.052491 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.056299 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.056544 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.056873 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.057206 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.059075 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.059575 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.068208 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsbld\" (UniqueName: \"kubernetes.io/projected/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-kube-api-access-hsbld\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.170624 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.301444 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.360420 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-catalog-content\") pod \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.360548 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mskvt\" (UniqueName: \"kubernetes.io/projected/ef15b0b6-7e6e-4a09-93ad-64137cee3543-kube-api-access-mskvt\") pod \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.360746 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-utilities\") pod \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\" (UID: \"ef15b0b6-7e6e-4a09-93ad-64137cee3543\") " Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.364012 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-utilities" (OuterVolumeSpecName: "utilities") pod "ef15b0b6-7e6e-4a09-93ad-64137cee3543" (UID: "ef15b0b6-7e6e-4a09-93ad-64137cee3543"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.367665 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef15b0b6-7e6e-4a09-93ad-64137cee3543-kube-api-access-mskvt" (OuterVolumeSpecName: "kube-api-access-mskvt") pod "ef15b0b6-7e6e-4a09-93ad-64137cee3543" (UID: "ef15b0b6-7e6e-4a09-93ad-64137cee3543"). InnerVolumeSpecName "kube-api-access-mskvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.382156 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef15b0b6-7e6e-4a09-93ad-64137cee3543" (UID: "ef15b0b6-7e6e-4a09-93ad-64137cee3543"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.464354 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.464407 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef15b0b6-7e6e-4a09-93ad-64137cee3543-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.464425 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mskvt\" (UniqueName: \"kubernetes.io/projected/ef15b0b6-7e6e-4a09-93ad-64137cee3543-kube-api-access-mskvt\") on node \"crc\" DevicePath \"\"" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.737866 4735 generic.go:334] "Generic (PLEG): container finished" podID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerID="731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721" exitCode=0 Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.738166 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzsp" event={"ID":"ef15b0b6-7e6e-4a09-93ad-64137cee3543","Type":"ContainerDied","Data":"731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721"} Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.738201 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzsp" event={"ID":"ef15b0b6-7e6e-4a09-93ad-64137cee3543","Type":"ContainerDied","Data":"c65fd177e4c0b0a1998fbc112612b2791b3aea267a1e61909e600846f3755b78"} Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.738220 4735 scope.go:117] "RemoveContainer" containerID="731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.738363 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfzsp" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.779174 4735 scope.go:117] "RemoveContainer" containerID="5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.780434 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzsp"] Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.793541 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzsp"] Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.802361 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl"] Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.821854 4735 scope.go:117] "RemoveContainer" containerID="1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.870580 4735 scope.go:117] "RemoveContainer" containerID="731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721" Nov 22 08:50:21 crc kubenswrapper[4735]: E1122 08:50:21.872214 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721\": container with ID starting with 731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721 not found: ID does not exist" containerID="731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.872262 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721"} err="failed to get container status \"731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721\": rpc error: code = NotFound desc = could not find container \"731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721\": container with ID starting with 731913d254c62b1cc38656843b2d0cd536a4c7f672fbd08f2bd68e491a84c721 not found: ID does not exist" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.872299 4735 scope.go:117] "RemoveContainer" containerID="5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f" Nov 22 08:50:21 crc kubenswrapper[4735]: E1122 08:50:21.872921 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f\": container with ID starting with 5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f not found: ID does not exist" containerID="5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.872981 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f"} err="failed to get container status \"5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f\": rpc error: code = NotFound desc = could not find container \"5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f\": container with ID starting with 5b8537ac177309af391ea216c2a3c64c7a54ea77de57772f698d68dd8a9d615f not found: ID does not exist" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.873022 4735 scope.go:117] "RemoveContainer" containerID="1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156" Nov 22 08:50:21 crc kubenswrapper[4735]: E1122 08:50:21.873387 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156\": container with ID starting with 1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156 not found: ID does not exist" containerID="1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156" Nov 22 08:50:21 crc kubenswrapper[4735]: I1122 08:50:21.873430 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156"} err="failed to get container status \"1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156\": rpc error: code = NotFound desc = could not find container \"1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156\": container with ID starting with 1a85e1781f5e8c92f17d4c6cac917dcfbe6fa77044e2fe8fd7ad9ca459530156 not found: ID does not exist" Nov 22 08:50:22 crc kubenswrapper[4735]: I1122 08:50:22.750704 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" event={"ID":"4d1a66ad-914a-4d94-ba26-7ed039cadf8f","Type":"ContainerStarted","Data":"248aa3af98c33b009de78b5d669116ebbec04b0374fa70a9f65078782bc91e2a"} Nov 22 08:50:22 crc kubenswrapper[4735]: I1122 08:50:22.751065 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" event={"ID":"4d1a66ad-914a-4d94-ba26-7ed039cadf8f","Type":"ContainerStarted","Data":"583eed7d72e05958af00046b56905f87b362e808c33d7b6bf85d418856b54df5"} Nov 22 08:50:22 crc kubenswrapper[4735]: I1122 08:50:22.795420 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" podStartSLOduration=2.323143912 podStartE2EDuration="2.795399032s" podCreationTimestamp="2025-11-22 08:50:20 +0000 UTC" firstStartedPulling="2025-11-22 08:50:21.804859822 +0000 UTC m=+2843.409198437" lastFinishedPulling="2025-11-22 08:50:22.277114962 +0000 UTC m=+2843.881453557" observedRunningTime="2025-11-22 08:50:22.78514667 +0000 UTC m=+2844.389485305" watchObservedRunningTime="2025-11-22 08:50:22.795399032 +0000 UTC m=+2844.399737637" Nov 22 08:50:23 crc kubenswrapper[4735]: I1122 08:50:23.279320 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" path="/var/lib/kubelet/pods/ef15b0b6-7e6e-4a09-93ad-64137cee3543/volumes" Nov 22 08:50:46 crc kubenswrapper[4735]: I1122 08:50:46.131566 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:50:46 crc kubenswrapper[4735]: I1122 08:50:46.133652 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:51:16 crc kubenswrapper[4735]: I1122 08:51:16.131029 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:51:16 crc kubenswrapper[4735]: I1122 08:51:16.131623 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.131728 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.133677 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.133898 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.135028 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.135226 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" gracePeriod=600 Nov 22 08:51:46 crc kubenswrapper[4735]: E1122 08:51:46.255687 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.765202 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" exitCode=0 Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.765661 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2"} Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.765710 4735 scope.go:117] "RemoveContainer" containerID="fd09bda8cc7301147359675a9e8663a853932b4858c0e3f7f11321068fb04ea8" Nov 22 08:51:46 crc kubenswrapper[4735]: I1122 08:51:46.766824 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:51:46 crc kubenswrapper[4735]: E1122 08:51:46.767605 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:51:58 crc kubenswrapper[4735]: I1122 08:51:58.263475 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:51:58 crc kubenswrapper[4735]: E1122 08:51:58.264274 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:52:10 crc kubenswrapper[4735]: I1122 08:52:10.263402 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:52:10 crc kubenswrapper[4735]: E1122 08:52:10.264248 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:52:23 crc kubenswrapper[4735]: I1122 08:52:23.263040 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:52:23 crc kubenswrapper[4735]: E1122 08:52:23.265802 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:52:34 crc kubenswrapper[4735]: I1122 08:52:34.264333 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:52:34 crc kubenswrapper[4735]: E1122 08:52:34.265841 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.625264 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vdcxg"] Nov 22 08:52:46 crc kubenswrapper[4735]: E1122 08:52:46.626883 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerName="extract-content" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.626917 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerName="extract-content" Nov 22 08:52:46 crc kubenswrapper[4735]: E1122 08:52:46.626940 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerName="registry-server" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.626957 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerName="registry-server" Nov 22 08:52:46 crc kubenswrapper[4735]: E1122 08:52:46.627062 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerName="extract-utilities" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.627080 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerName="extract-utilities" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.627688 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef15b0b6-7e6e-4a09-93ad-64137cee3543" containerName="registry-server" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.631409 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.643608 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vdcxg"] Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.738326 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtxzp\" (UniqueName: \"kubernetes.io/projected/c9c9e39d-9395-4764-8812-be8bfefb130d-kube-api-access-gtxzp\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.738569 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-utilities\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.738598 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-catalog-content\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.841088 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-utilities\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.841148 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-catalog-content\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.841284 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtxzp\" (UniqueName: \"kubernetes.io/projected/c9c9e39d-9395-4764-8812-be8bfefb130d-kube-api-access-gtxzp\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.841846 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-utilities\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.841896 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-catalog-content\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.862329 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtxzp\" (UniqueName: \"kubernetes.io/projected/c9c9e39d-9395-4764-8812-be8bfefb130d-kube-api-access-gtxzp\") pod \"community-operators-vdcxg\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:46 crc kubenswrapper[4735]: I1122 08:52:46.964897 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:47 crc kubenswrapper[4735]: I1122 08:52:47.499801 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vdcxg"] Nov 22 08:52:48 crc kubenswrapper[4735]: I1122 08:52:48.263877 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:52:48 crc kubenswrapper[4735]: E1122 08:52:48.264508 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:52:48 crc kubenswrapper[4735]: I1122 08:52:48.497517 4735 generic.go:334] "Generic (PLEG): container finished" podID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerID="2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65" exitCode=0 Nov 22 08:52:48 crc kubenswrapper[4735]: I1122 08:52:48.497647 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdcxg" event={"ID":"c9c9e39d-9395-4764-8812-be8bfefb130d","Type":"ContainerDied","Data":"2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65"} Nov 22 08:52:48 crc kubenswrapper[4735]: I1122 08:52:48.498164 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdcxg" event={"ID":"c9c9e39d-9395-4764-8812-be8bfefb130d","Type":"ContainerStarted","Data":"63a3c67c44f2524778a84358f5e0b08f5d8e1e806f873971e26b591ab4df93d7"} Nov 22 08:52:48 crc kubenswrapper[4735]: I1122 08:52:48.502115 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:52:50 crc kubenswrapper[4735]: I1122 08:52:50.523056 4735 generic.go:334] "Generic (PLEG): container finished" podID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerID="3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115" exitCode=0 Nov 22 08:52:50 crc kubenswrapper[4735]: I1122 08:52:50.523153 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdcxg" event={"ID":"c9c9e39d-9395-4764-8812-be8bfefb130d","Type":"ContainerDied","Data":"3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115"} Nov 22 08:52:51 crc kubenswrapper[4735]: I1122 08:52:51.536691 4735 generic.go:334] "Generic (PLEG): container finished" podID="4d1a66ad-914a-4d94-ba26-7ed039cadf8f" containerID="248aa3af98c33b009de78b5d669116ebbec04b0374fa70a9f65078782bc91e2a" exitCode=0 Nov 22 08:52:51 crc kubenswrapper[4735]: I1122 08:52:51.536758 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" event={"ID":"4d1a66ad-914a-4d94-ba26-7ed039cadf8f","Type":"ContainerDied","Data":"248aa3af98c33b009de78b5d669116ebbec04b0374fa70a9f65078782bc91e2a"} Nov 22 08:52:51 crc kubenswrapper[4735]: I1122 08:52:51.540315 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdcxg" event={"ID":"c9c9e39d-9395-4764-8812-be8bfefb130d","Type":"ContainerStarted","Data":"4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a"} Nov 22 08:52:51 crc kubenswrapper[4735]: I1122 08:52:51.594500 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vdcxg" podStartSLOduration=3.164748524 podStartE2EDuration="5.594481935s" podCreationTimestamp="2025-11-22 08:52:46 +0000 UTC" firstStartedPulling="2025-11-22 08:52:48.501492084 +0000 UTC m=+2990.105830699" lastFinishedPulling="2025-11-22 08:52:50.931225485 +0000 UTC m=+2992.535564110" observedRunningTime="2025-11-22 08:52:51.585779544 +0000 UTC m=+2993.190118149" watchObservedRunningTime="2025-11-22 08:52:51.594481935 +0000 UTC m=+2993.198820540" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.202008 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-crr5r"] Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.205296 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.219632 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-crr5r"] Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.318787 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-catalog-content\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.318898 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-utilities\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.319018 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf4n5\" (UniqueName: \"kubernetes.io/projected/03f0db8b-0c08-4911-8b4b-0fbc571824aa-kube-api-access-jf4n5\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.422039 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-catalog-content\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.422337 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-utilities\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.422386 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf4n5\" (UniqueName: \"kubernetes.io/projected/03f0db8b-0c08-4911-8b4b-0fbc571824aa-kube-api-access-jf4n5\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.423251 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-catalog-content\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.423486 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-utilities\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.450588 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf4n5\" (UniqueName: \"kubernetes.io/projected/03f0db8b-0c08-4911-8b4b-0fbc571824aa-kube-api-access-jf4n5\") pod \"redhat-operators-crr5r\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:52 crc kubenswrapper[4735]: I1122 08:52:52.570172 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.121052 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:52:53 crc kubenswrapper[4735]: W1122 08:52:53.172560 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03f0db8b_0c08_4911_8b4b_0fbc571824aa.slice/crio-0db0e7899fa25bb73da734fcf4230112b58514cd05efa3dc42d7217baf02d894 WatchSource:0}: Error finding container 0db0e7899fa25bb73da734fcf4230112b58514cd05efa3dc42d7217baf02d894: Status 404 returned error can't find the container with id 0db0e7899fa25bb73da734fcf4230112b58514cd05efa3dc42d7217baf02d894 Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.175953 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-crr5r"] Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.241322 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-inventory\") pod \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.241523 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsbld\" (UniqueName: \"kubernetes.io/projected/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-kube-api-access-hsbld\") pod \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.241553 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ssh-key\") pod \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.241584 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-0\") pod \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.241751 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-telemetry-combined-ca-bundle\") pod \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.241897 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-2\") pod \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.242323 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-1\") pod \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\" (UID: \"4d1a66ad-914a-4d94-ba26-7ed039cadf8f\") " Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.246734 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "4d1a66ad-914a-4d94-ba26-7ed039cadf8f" (UID: "4d1a66ad-914a-4d94-ba26-7ed039cadf8f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.256191 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-kube-api-access-hsbld" (OuterVolumeSpecName: "kube-api-access-hsbld") pod "4d1a66ad-914a-4d94-ba26-7ed039cadf8f" (UID: "4d1a66ad-914a-4d94-ba26-7ed039cadf8f"). InnerVolumeSpecName "kube-api-access-hsbld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.281682 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d1a66ad-914a-4d94-ba26-7ed039cadf8f" (UID: "4d1a66ad-914a-4d94-ba26-7ed039cadf8f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.287225 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "4d1a66ad-914a-4d94-ba26-7ed039cadf8f" (UID: "4d1a66ad-914a-4d94-ba26-7ed039cadf8f"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.293377 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "4d1a66ad-914a-4d94-ba26-7ed039cadf8f" (UID: "4d1a66ad-914a-4d94-ba26-7ed039cadf8f"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.295697 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-inventory" (OuterVolumeSpecName: "inventory") pod "4d1a66ad-914a-4d94-ba26-7ed039cadf8f" (UID: "4d1a66ad-914a-4d94-ba26-7ed039cadf8f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.299492 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "4d1a66ad-914a-4d94-ba26-7ed039cadf8f" (UID: "4d1a66ad-914a-4d94-ba26-7ed039cadf8f"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.346384 4735 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.347292 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.347329 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.347346 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.347357 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsbld\" (UniqueName: \"kubernetes.io/projected/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-kube-api-access-hsbld\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.347387 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.347398 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d1a66ad-914a-4d94-ba26-7ed039cadf8f-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.560964 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" event={"ID":"4d1a66ad-914a-4d94-ba26-7ed039cadf8f","Type":"ContainerDied","Data":"583eed7d72e05958af00046b56905f87b362e808c33d7b6bf85d418856b54df5"} Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.561011 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="583eed7d72e05958af00046b56905f87b362e808c33d7b6bf85d418856b54df5" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.561072 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.563845 4735 generic.go:334] "Generic (PLEG): container finished" podID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerID="62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24" exitCode=0 Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.563894 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crr5r" event={"ID":"03f0db8b-0c08-4911-8b4b-0fbc571824aa","Type":"ContainerDied","Data":"62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24"} Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.563952 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crr5r" event={"ID":"03f0db8b-0c08-4911-8b4b-0fbc571824aa","Type":"ContainerStarted","Data":"0db0e7899fa25bb73da734fcf4230112b58514cd05efa3dc42d7217baf02d894"} Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.714595 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r"] Nov 22 08:52:53 crc kubenswrapper[4735]: E1122 08:52:53.715408 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d1a66ad-914a-4d94-ba26-7ed039cadf8f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.715422 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d1a66ad-914a-4d94-ba26-7ed039cadf8f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.715708 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d1a66ad-914a-4d94-ba26-7ed039cadf8f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.716637 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.720436 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.720730 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.721174 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.721207 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.721689 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.749623 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r"] Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.776910 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.777224 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.777358 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.777523 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.777638 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.777756 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z69x8\" (UniqueName: \"kubernetes.io/projected/a393bb4a-7413-4a36-bc56-e6a18f17fffa-kube-api-access-z69x8\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.777839 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.880187 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.880366 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.880597 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.880719 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.880848 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z69x8\" (UniqueName: \"kubernetes.io/projected/a393bb4a-7413-4a36-bc56-e6a18f17fffa-kube-api-access-z69x8\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.881428 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.882095 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.884991 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.885003 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.885289 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.886820 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.887318 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.887378 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:53 crc kubenswrapper[4735]: I1122 08:52:53.896246 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z69x8\" (UniqueName: \"kubernetes.io/projected/a393bb4a-7413-4a36-bc56-e6a18f17fffa-kube-api-access-z69x8\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:54 crc kubenswrapper[4735]: I1122 08:52:54.040938 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:52:54 crc kubenswrapper[4735]: W1122 08:52:54.662968 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda393bb4a_7413_4a36_bc56_e6a18f17fffa.slice/crio-b4cce008a240fbdeafea93976c8c63a20b1ad147d9ad279d58fc639c9ba1c5fe WatchSource:0}: Error finding container b4cce008a240fbdeafea93976c8c63a20b1ad147d9ad279d58fc639c9ba1c5fe: Status 404 returned error can't find the container with id b4cce008a240fbdeafea93976c8c63a20b1ad147d9ad279d58fc639c9ba1c5fe Nov 22 08:52:54 crc kubenswrapper[4735]: I1122 08:52:54.663260 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r"] Nov 22 08:52:55 crc kubenswrapper[4735]: I1122 08:52:55.601509 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" event={"ID":"a393bb4a-7413-4a36-bc56-e6a18f17fffa","Type":"ContainerStarted","Data":"9b0da910ce6992905fad1345d8e3f1b44219ef36c2138db4699d4e0d30c21f97"} Nov 22 08:52:55 crc kubenswrapper[4735]: I1122 08:52:55.602062 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" event={"ID":"a393bb4a-7413-4a36-bc56-e6a18f17fffa","Type":"ContainerStarted","Data":"b4cce008a240fbdeafea93976c8c63a20b1ad147d9ad279d58fc639c9ba1c5fe"} Nov 22 08:52:55 crc kubenswrapper[4735]: I1122 08:52:55.605648 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crr5r" event={"ID":"03f0db8b-0c08-4911-8b4b-0fbc571824aa","Type":"ContainerStarted","Data":"139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c"} Nov 22 08:52:56 crc kubenswrapper[4735]: I1122 08:52:56.649295 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" podStartSLOduration=3.191705668 podStartE2EDuration="3.649262864s" podCreationTimestamp="2025-11-22 08:52:53 +0000 UTC" firstStartedPulling="2025-11-22 08:52:54.674945239 +0000 UTC m=+2996.279283864" lastFinishedPulling="2025-11-22 08:52:55.132502445 +0000 UTC m=+2996.736841060" observedRunningTime="2025-11-22 08:52:56.642915209 +0000 UTC m=+2998.247253854" watchObservedRunningTime="2025-11-22 08:52:56.649262864 +0000 UTC m=+2998.253601499" Nov 22 08:52:56 crc kubenswrapper[4735]: I1122 08:52:56.966779 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:56 crc kubenswrapper[4735]: I1122 08:52:56.967714 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:57 crc kubenswrapper[4735]: I1122 08:52:57.015376 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:57 crc kubenswrapper[4735]: I1122 08:52:57.751707 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:52:58 crc kubenswrapper[4735]: I1122 08:52:58.180282 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vdcxg"] Nov 22 08:52:59 crc kubenswrapper[4735]: I1122 08:52:59.688057 4735 generic.go:334] "Generic (PLEG): container finished" podID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerID="139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c" exitCode=0 Nov 22 08:52:59 crc kubenswrapper[4735]: I1122 08:52:59.688136 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crr5r" event={"ID":"03f0db8b-0c08-4911-8b4b-0fbc571824aa","Type":"ContainerDied","Data":"139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c"} Nov 22 08:52:59 crc kubenswrapper[4735]: I1122 08:52:59.688625 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vdcxg" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerName="registry-server" containerID="cri-o://4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a" gracePeriod=2 Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.289253 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.363911 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-utilities\") pod \"c9c9e39d-9395-4764-8812-be8bfefb130d\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.364201 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtxzp\" (UniqueName: \"kubernetes.io/projected/c9c9e39d-9395-4764-8812-be8bfefb130d-kube-api-access-gtxzp\") pod \"c9c9e39d-9395-4764-8812-be8bfefb130d\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.364567 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-catalog-content\") pod \"c9c9e39d-9395-4764-8812-be8bfefb130d\" (UID: \"c9c9e39d-9395-4764-8812-be8bfefb130d\") " Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.364684 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-utilities" (OuterVolumeSpecName: "utilities") pod "c9c9e39d-9395-4764-8812-be8bfefb130d" (UID: "c9c9e39d-9395-4764-8812-be8bfefb130d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.365406 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.371166 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c9e39d-9395-4764-8812-be8bfefb130d-kube-api-access-gtxzp" (OuterVolumeSpecName: "kube-api-access-gtxzp") pod "c9c9e39d-9395-4764-8812-be8bfefb130d" (UID: "c9c9e39d-9395-4764-8812-be8bfefb130d"). InnerVolumeSpecName "kube-api-access-gtxzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.410976 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9c9e39d-9395-4764-8812-be8bfefb130d" (UID: "c9c9e39d-9395-4764-8812-be8bfefb130d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.467482 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c9e39d-9395-4764-8812-be8bfefb130d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.467521 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtxzp\" (UniqueName: \"kubernetes.io/projected/c9c9e39d-9395-4764-8812-be8bfefb130d-kube-api-access-gtxzp\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.729846 4735 generic.go:334] "Generic (PLEG): container finished" podID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerID="4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a" exitCode=0 Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.729909 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdcxg" event={"ID":"c9c9e39d-9395-4764-8812-be8bfefb130d","Type":"ContainerDied","Data":"4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a"} Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.729953 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdcxg" event={"ID":"c9c9e39d-9395-4764-8812-be8bfefb130d","Type":"ContainerDied","Data":"63a3c67c44f2524778a84358f5e0b08f5d8e1e806f873971e26b591ab4df93d7"} Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.729990 4735 scope.go:117] "RemoveContainer" containerID="4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.730281 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdcxg" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.770497 4735 scope.go:117] "RemoveContainer" containerID="3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.786133 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vdcxg"] Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.798252 4735 scope.go:117] "RemoveContainer" containerID="2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.798330 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vdcxg"] Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.872161 4735 scope.go:117] "RemoveContainer" containerID="4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a" Nov 22 08:53:00 crc kubenswrapper[4735]: E1122 08:53:00.872746 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a\": container with ID starting with 4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a not found: ID does not exist" containerID="4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.872802 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a"} err="failed to get container status \"4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a\": rpc error: code = NotFound desc = could not find container \"4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a\": container with ID starting with 4b66f05dab3e9092c4cbcb4a2e8d695c456ffc4ccef5374be83c0a2b1f94517a not found: ID does not exist" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.872842 4735 scope.go:117] "RemoveContainer" containerID="3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115" Nov 22 08:53:00 crc kubenswrapper[4735]: E1122 08:53:00.873343 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115\": container with ID starting with 3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115 not found: ID does not exist" containerID="3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.873386 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115"} err="failed to get container status \"3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115\": rpc error: code = NotFound desc = could not find container \"3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115\": container with ID starting with 3af65163325d6abe2bdc27ec8cf3255ddc191276bd764d52e0ffc12aac2c1115 not found: ID does not exist" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.873413 4735 scope.go:117] "RemoveContainer" containerID="2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65" Nov 22 08:53:00 crc kubenswrapper[4735]: E1122 08:53:00.873839 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65\": container with ID starting with 2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65 not found: ID does not exist" containerID="2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65" Nov 22 08:53:00 crc kubenswrapper[4735]: I1122 08:53:00.873870 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65"} err="failed to get container status \"2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65\": rpc error: code = NotFound desc = could not find container \"2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65\": container with ID starting with 2d7246e852ba04aa6d5ce3ba7c99de675be05deaf8482fa165a538582af4ec65 not found: ID does not exist" Nov 22 08:53:01 crc kubenswrapper[4735]: I1122 08:53:01.278209 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" path="/var/lib/kubelet/pods/c9c9e39d-9395-4764-8812-be8bfefb130d/volumes" Nov 22 08:53:01 crc kubenswrapper[4735]: I1122 08:53:01.741147 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crr5r" event={"ID":"03f0db8b-0c08-4911-8b4b-0fbc571824aa","Type":"ContainerStarted","Data":"3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218"} Nov 22 08:53:01 crc kubenswrapper[4735]: I1122 08:53:01.761730 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-crr5r" podStartSLOduration=2.861406324 podStartE2EDuration="9.761708295s" podCreationTimestamp="2025-11-22 08:52:52 +0000 UTC" firstStartedPulling="2025-11-22 08:52:53.566240318 +0000 UTC m=+2995.170578923" lastFinishedPulling="2025-11-22 08:53:00.466542289 +0000 UTC m=+3002.070880894" observedRunningTime="2025-11-22 08:53:01.755240917 +0000 UTC m=+3003.359579542" watchObservedRunningTime="2025-11-22 08:53:01.761708295 +0000 UTC m=+3003.366046900" Nov 22 08:53:02 crc kubenswrapper[4735]: I1122 08:53:02.570587 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:53:02 crc kubenswrapper[4735]: I1122 08:53:02.570648 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:53:03 crc kubenswrapper[4735]: I1122 08:53:03.264164 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:53:03 crc kubenswrapper[4735]: E1122 08:53:03.264904 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:53:03 crc kubenswrapper[4735]: I1122 08:53:03.620163 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-crr5r" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="registry-server" probeResult="failure" output=< Nov 22 08:53:03 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 08:53:03 crc kubenswrapper[4735]: > Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.320892 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t5q7z"] Nov 22 08:53:11 crc kubenswrapper[4735]: E1122 08:53:11.322685 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerName="registry-server" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.322703 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerName="registry-server" Nov 22 08:53:11 crc kubenswrapper[4735]: E1122 08:53:11.322755 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerName="extract-utilities" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.322764 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerName="extract-utilities" Nov 22 08:53:11 crc kubenswrapper[4735]: E1122 08:53:11.322785 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerName="extract-content" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.322792 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerName="extract-content" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.323039 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c9e39d-9395-4764-8812-be8bfefb130d" containerName="registry-server" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.324942 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.336334 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t5q7z"] Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.358364 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-utilities\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.358576 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s82hf\" (UniqueName: \"kubernetes.io/projected/691ac120-c31b-4133-9847-4bb191808eb8-kube-api-access-s82hf\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.358706 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-catalog-content\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.461712 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-utilities\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.462019 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s82hf\" (UniqueName: \"kubernetes.io/projected/691ac120-c31b-4133-9847-4bb191808eb8-kube-api-access-s82hf\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.462122 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-catalog-content\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.462251 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-utilities\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.462495 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-catalog-content\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.488377 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s82hf\" (UniqueName: \"kubernetes.io/projected/691ac120-c31b-4133-9847-4bb191808eb8-kube-api-access-s82hf\") pod \"certified-operators-t5q7z\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:11 crc kubenswrapper[4735]: I1122 08:53:11.656138 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:12 crc kubenswrapper[4735]: W1122 08:53:12.317575 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod691ac120_c31b_4133_9847_4bb191808eb8.slice/crio-032a146ab96beb37f9c3a5445b5a20e6fcc3f59197d08a2eac0baa7d4ea9a81e WatchSource:0}: Error finding container 032a146ab96beb37f9c3a5445b5a20e6fcc3f59197d08a2eac0baa7d4ea9a81e: Status 404 returned error can't find the container with id 032a146ab96beb37f9c3a5445b5a20e6fcc3f59197d08a2eac0baa7d4ea9a81e Nov 22 08:53:12 crc kubenswrapper[4735]: I1122 08:53:12.333320 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t5q7z"] Nov 22 08:53:12 crc kubenswrapper[4735]: I1122 08:53:12.623175 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:53:12 crc kubenswrapper[4735]: I1122 08:53:12.673508 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:53:12 crc kubenswrapper[4735]: I1122 08:53:12.887905 4735 generic.go:334] "Generic (PLEG): container finished" podID="691ac120-c31b-4133-9847-4bb191808eb8" containerID="f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2" exitCode=0 Nov 22 08:53:12 crc kubenswrapper[4735]: I1122 08:53:12.888092 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t5q7z" event={"ID":"691ac120-c31b-4133-9847-4bb191808eb8","Type":"ContainerDied","Data":"f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2"} Nov 22 08:53:12 crc kubenswrapper[4735]: I1122 08:53:12.888240 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t5q7z" event={"ID":"691ac120-c31b-4133-9847-4bb191808eb8","Type":"ContainerStarted","Data":"032a146ab96beb37f9c3a5445b5a20e6fcc3f59197d08a2eac0baa7d4ea9a81e"} Nov 22 08:53:13 crc kubenswrapper[4735]: I1122 08:53:13.904227 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t5q7z" event={"ID":"691ac120-c31b-4133-9847-4bb191808eb8","Type":"ContainerStarted","Data":"06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2"} Nov 22 08:53:14 crc kubenswrapper[4735]: I1122 08:53:14.263349 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:53:14 crc kubenswrapper[4735]: E1122 08:53:14.263655 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:53:14 crc kubenswrapper[4735]: I1122 08:53:14.712416 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-crr5r"] Nov 22 08:53:14 crc kubenswrapper[4735]: I1122 08:53:14.714055 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-crr5r" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="registry-server" containerID="cri-o://3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218" gracePeriod=2 Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.831794 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.876751 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-catalog-content\") pod \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.876938 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf4n5\" (UniqueName: \"kubernetes.io/projected/03f0db8b-0c08-4911-8b4b-0fbc571824aa-kube-api-access-jf4n5\") pod \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.877029 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-utilities\") pod \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\" (UID: \"03f0db8b-0c08-4911-8b4b-0fbc571824aa\") " Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.878635 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-utilities" (OuterVolumeSpecName: "utilities") pod "03f0db8b-0c08-4911-8b4b-0fbc571824aa" (UID: "03f0db8b-0c08-4911-8b4b-0fbc571824aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.888803 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03f0db8b-0c08-4911-8b4b-0fbc571824aa-kube-api-access-jf4n5" (OuterVolumeSpecName: "kube-api-access-jf4n5") pod "03f0db8b-0c08-4911-8b4b-0fbc571824aa" (UID: "03f0db8b-0c08-4911-8b4b-0fbc571824aa"). InnerVolumeSpecName "kube-api-access-jf4n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.931266 4735 generic.go:334] "Generic (PLEG): container finished" podID="691ac120-c31b-4133-9847-4bb191808eb8" containerID="06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2" exitCode=0 Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.931352 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t5q7z" event={"ID":"691ac120-c31b-4133-9847-4bb191808eb8","Type":"ContainerDied","Data":"06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2"} Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.940999 4735 generic.go:334] "Generic (PLEG): container finished" podID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerID="3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218" exitCode=0 Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.941044 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crr5r" event={"ID":"03f0db8b-0c08-4911-8b4b-0fbc571824aa","Type":"ContainerDied","Data":"3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218"} Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.941072 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crr5r" event={"ID":"03f0db8b-0c08-4911-8b4b-0fbc571824aa","Type":"ContainerDied","Data":"0db0e7899fa25bb73da734fcf4230112b58514cd05efa3dc42d7217baf02d894"} Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.941089 4735 scope.go:117] "RemoveContainer" containerID="3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218" Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.941270 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-crr5r" Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.979907 4735 scope.go:117] "RemoveContainer" containerID="139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c" Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.982256 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf4n5\" (UniqueName: \"kubernetes.io/projected/03f0db8b-0c08-4911-8b4b-0fbc571824aa-kube-api-access-jf4n5\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.982284 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:15 crc kubenswrapper[4735]: I1122 08:53:15.994182 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03f0db8b-0c08-4911-8b4b-0fbc571824aa" (UID: "03f0db8b-0c08-4911-8b4b-0fbc571824aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.004795 4735 scope.go:117] "RemoveContainer" containerID="62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.053769 4735 scope.go:117] "RemoveContainer" containerID="3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218" Nov 22 08:53:16 crc kubenswrapper[4735]: E1122 08:53:16.054440 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218\": container with ID starting with 3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218 not found: ID does not exist" containerID="3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.054499 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218"} err="failed to get container status \"3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218\": rpc error: code = NotFound desc = could not find container \"3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218\": container with ID starting with 3f0c790a5540d4dab38e039e00176494659a65261bb7142c65278f3a3a0e7218 not found: ID does not exist" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.054528 4735 scope.go:117] "RemoveContainer" containerID="139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c" Nov 22 08:53:16 crc kubenswrapper[4735]: E1122 08:53:16.054943 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c\": container with ID starting with 139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c not found: ID does not exist" containerID="139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.054970 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c"} err="failed to get container status \"139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c\": rpc error: code = NotFound desc = could not find container \"139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c\": container with ID starting with 139c690fa9e0cbea8ed7a22e131e57ada3b10a7eb375a3a59110966c19be812c not found: ID does not exist" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.054989 4735 scope.go:117] "RemoveContainer" containerID="62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24" Nov 22 08:53:16 crc kubenswrapper[4735]: E1122 08:53:16.055492 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24\": container with ID starting with 62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24 not found: ID does not exist" containerID="62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.055547 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24"} err="failed to get container status \"62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24\": rpc error: code = NotFound desc = could not find container \"62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24\": container with ID starting with 62006695301f4d26e52095e9d0b49b13353e4e89a0acaf24a2a96c29f2e7fb24 not found: ID does not exist" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.084744 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03f0db8b-0c08-4911-8b4b-0fbc571824aa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.280704 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-crr5r"] Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.290113 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-crr5r"] Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.960093 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t5q7z" event={"ID":"691ac120-c31b-4133-9847-4bb191808eb8","Type":"ContainerStarted","Data":"15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287"} Nov 22 08:53:16 crc kubenswrapper[4735]: I1122 08:53:16.983187 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t5q7z" podStartSLOduration=2.489885961 podStartE2EDuration="5.983169356s" podCreationTimestamp="2025-11-22 08:53:11 +0000 UTC" firstStartedPulling="2025-11-22 08:53:12.891165002 +0000 UTC m=+3014.495503607" lastFinishedPulling="2025-11-22 08:53:16.384448357 +0000 UTC m=+3017.988787002" observedRunningTime="2025-11-22 08:53:16.97713564 +0000 UTC m=+3018.581474245" watchObservedRunningTime="2025-11-22 08:53:16.983169356 +0000 UTC m=+3018.587507961" Nov 22 08:53:17 crc kubenswrapper[4735]: I1122 08:53:17.283428 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" path="/var/lib/kubelet/pods/03f0db8b-0c08-4911-8b4b-0fbc571824aa/volumes" Nov 22 08:53:21 crc kubenswrapper[4735]: I1122 08:53:21.656475 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:21 crc kubenswrapper[4735]: I1122 08:53:21.657010 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:21 crc kubenswrapper[4735]: I1122 08:53:21.735863 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:22 crc kubenswrapper[4735]: I1122 08:53:22.071956 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:22 crc kubenswrapper[4735]: I1122 08:53:22.318256 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t5q7z"] Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.051818 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t5q7z" podUID="691ac120-c31b-4133-9847-4bb191808eb8" containerName="registry-server" containerID="cri-o://15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287" gracePeriod=2 Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.577652 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.690841 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-catalog-content\") pod \"691ac120-c31b-4133-9847-4bb191808eb8\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.691087 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-utilities\") pod \"691ac120-c31b-4133-9847-4bb191808eb8\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.691136 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s82hf\" (UniqueName: \"kubernetes.io/projected/691ac120-c31b-4133-9847-4bb191808eb8-kube-api-access-s82hf\") pod \"691ac120-c31b-4133-9847-4bb191808eb8\" (UID: \"691ac120-c31b-4133-9847-4bb191808eb8\") " Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.691778 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-utilities" (OuterVolumeSpecName: "utilities") pod "691ac120-c31b-4133-9847-4bb191808eb8" (UID: "691ac120-c31b-4133-9847-4bb191808eb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.705778 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/691ac120-c31b-4133-9847-4bb191808eb8-kube-api-access-s82hf" (OuterVolumeSpecName: "kube-api-access-s82hf") pod "691ac120-c31b-4133-9847-4bb191808eb8" (UID: "691ac120-c31b-4133-9847-4bb191808eb8"). InnerVolumeSpecName "kube-api-access-s82hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.736852 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "691ac120-c31b-4133-9847-4bb191808eb8" (UID: "691ac120-c31b-4133-9847-4bb191808eb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.794261 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.794490 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/691ac120-c31b-4133-9847-4bb191808eb8-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:24 crc kubenswrapper[4735]: I1122 08:53:24.794605 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s82hf\" (UniqueName: \"kubernetes.io/projected/691ac120-c31b-4133-9847-4bb191808eb8-kube-api-access-s82hf\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.067525 4735 generic.go:334] "Generic (PLEG): container finished" podID="691ac120-c31b-4133-9847-4bb191808eb8" containerID="15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287" exitCode=0 Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.067604 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t5q7z" event={"ID":"691ac120-c31b-4133-9847-4bb191808eb8","Type":"ContainerDied","Data":"15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287"} Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.068693 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t5q7z" event={"ID":"691ac120-c31b-4133-9847-4bb191808eb8","Type":"ContainerDied","Data":"032a146ab96beb37f9c3a5445b5a20e6fcc3f59197d08a2eac0baa7d4ea9a81e"} Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.067612 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t5q7z" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.068769 4735 scope.go:117] "RemoveContainer" containerID="15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.093454 4735 scope.go:117] "RemoveContainer" containerID="06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.121610 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t5q7z"] Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.130996 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t5q7z"] Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.134890 4735 scope.go:117] "RemoveContainer" containerID="f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.180206 4735 scope.go:117] "RemoveContainer" containerID="15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287" Nov 22 08:53:25 crc kubenswrapper[4735]: E1122 08:53:25.180882 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287\": container with ID starting with 15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287 not found: ID does not exist" containerID="15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.180996 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287"} err="failed to get container status \"15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287\": rpc error: code = NotFound desc = could not find container \"15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287\": container with ID starting with 15118425eb460815401a363ae4ff7dac4c8e225401c8b563c55dfd9a7f4d5287 not found: ID does not exist" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.181086 4735 scope.go:117] "RemoveContainer" containerID="06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2" Nov 22 08:53:25 crc kubenswrapper[4735]: E1122 08:53:25.181537 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2\": container with ID starting with 06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2 not found: ID does not exist" containerID="06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.181619 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2"} err="failed to get container status \"06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2\": rpc error: code = NotFound desc = could not find container \"06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2\": container with ID starting with 06ec5eb6abc2e8940df5662edc6ebeb56e08f9cdb0b1708f9f1c0175e59f54b2 not found: ID does not exist" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.181680 4735 scope.go:117] "RemoveContainer" containerID="f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2" Nov 22 08:53:25 crc kubenswrapper[4735]: E1122 08:53:25.182080 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2\": container with ID starting with f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2 not found: ID does not exist" containerID="f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.182171 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2"} err="failed to get container status \"f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2\": rpc error: code = NotFound desc = could not find container \"f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2\": container with ID starting with f995c3c5aec09494cdb45bed8512d85a4065990a1bb6949acb627fd1c9f50df2 not found: ID does not exist" Nov 22 08:53:25 crc kubenswrapper[4735]: I1122 08:53:25.279248 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="691ac120-c31b-4133-9847-4bb191808eb8" path="/var/lib/kubelet/pods/691ac120-c31b-4133-9847-4bb191808eb8/volumes" Nov 22 08:53:27 crc kubenswrapper[4735]: I1122 08:53:27.263994 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:53:27 crc kubenswrapper[4735]: E1122 08:53:27.265533 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:53:42 crc kubenswrapper[4735]: I1122 08:53:42.263826 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:53:42 crc kubenswrapper[4735]: E1122 08:53:42.265205 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:53:54 crc kubenswrapper[4735]: I1122 08:53:54.264218 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:53:54 crc kubenswrapper[4735]: E1122 08:53:54.265150 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:54:08 crc kubenswrapper[4735]: I1122 08:54:08.263979 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:54:08 crc kubenswrapper[4735]: E1122 08:54:08.264796 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:54:21 crc kubenswrapper[4735]: I1122 08:54:21.270226 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:54:21 crc kubenswrapper[4735]: E1122 08:54:21.271191 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:54:33 crc kubenswrapper[4735]: I1122 08:54:33.264088 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:54:33 crc kubenswrapper[4735]: E1122 08:54:33.265699 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:54:45 crc kubenswrapper[4735]: I1122 08:54:45.265235 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:54:45 crc kubenswrapper[4735]: E1122 08:54:45.266707 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:54:59 crc kubenswrapper[4735]: I1122 08:54:59.287295 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:54:59 crc kubenswrapper[4735]: E1122 08:54:59.288218 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:55:08 crc kubenswrapper[4735]: I1122 08:55:08.229630 4735 generic.go:334] "Generic (PLEG): container finished" podID="a393bb4a-7413-4a36-bc56-e6a18f17fffa" containerID="9b0da910ce6992905fad1345d8e3f1b44219ef36c2138db4699d4e0d30c21f97" exitCode=0 Nov 22 08:55:08 crc kubenswrapper[4735]: I1122 08:55:08.229754 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" event={"ID":"a393bb4a-7413-4a36-bc56-e6a18f17fffa","Type":"ContainerDied","Data":"9b0da910ce6992905fad1345d8e3f1b44219ef36c2138db4699d4e0d30c21f97"} Nov 22 08:55:08 crc kubenswrapper[4735]: E1122 08:55:08.232830 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda393bb4a_7413_4a36_bc56_e6a18f17fffa.slice/crio-9b0da910ce6992905fad1345d8e3f1b44219ef36c2138db4699d4e0d30c21f97.scope\": RecentStats: unable to find data in memory cache]" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.723006 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.804506 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-1\") pod \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.804752 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-0\") pod \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.804799 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-telemetry-power-monitoring-combined-ca-bundle\") pod \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.804870 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z69x8\" (UniqueName: \"kubernetes.io/projected/a393bb4a-7413-4a36-bc56-e6a18f17fffa-kube-api-access-z69x8\") pod \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.804903 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-2\") pod \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.804931 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-inventory\") pod \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.804946 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ssh-key\") pod \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\" (UID: \"a393bb4a-7413-4a36-bc56-e6a18f17fffa\") " Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.811745 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a393bb4a-7413-4a36-bc56-e6a18f17fffa-kube-api-access-z69x8" (OuterVolumeSpecName: "kube-api-access-z69x8") pod "a393bb4a-7413-4a36-bc56-e6a18f17fffa" (UID: "a393bb4a-7413-4a36-bc56-e6a18f17fffa"). InnerVolumeSpecName "kube-api-access-z69x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.828324 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "a393bb4a-7413-4a36-bc56-e6a18f17fffa" (UID: "a393bb4a-7413-4a36-bc56-e6a18f17fffa"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.841058 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "a393bb4a-7413-4a36-bc56-e6a18f17fffa" (UID: "a393bb4a-7413-4a36-bc56-e6a18f17fffa"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.842563 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-inventory" (OuterVolumeSpecName: "inventory") pod "a393bb4a-7413-4a36-bc56-e6a18f17fffa" (UID: "a393bb4a-7413-4a36-bc56-e6a18f17fffa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.843064 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "a393bb4a-7413-4a36-bc56-e6a18f17fffa" (UID: "a393bb4a-7413-4a36-bc56-e6a18f17fffa"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.845787 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a393bb4a-7413-4a36-bc56-e6a18f17fffa" (UID: "a393bb4a-7413-4a36-bc56-e6a18f17fffa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.862992 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "a393bb4a-7413-4a36-bc56-e6a18f17fffa" (UID: "a393bb4a-7413-4a36-bc56-e6a18f17fffa"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.907848 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.907889 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.907905 4735 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.907921 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z69x8\" (UniqueName: \"kubernetes.io/projected/a393bb4a-7413-4a36-bc56-e6a18f17fffa-kube-api-access-z69x8\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.907934 4735 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.907946 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:09 crc kubenswrapper[4735]: I1122 08:55:09.907958 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a393bb4a-7413-4a36-bc56-e6a18f17fffa-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.258141 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" event={"ID":"a393bb4a-7413-4a36-bc56-e6a18f17fffa","Type":"ContainerDied","Data":"b4cce008a240fbdeafea93976c8c63a20b1ad147d9ad279d58fc639c9ba1c5fe"} Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.258188 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4cce008a240fbdeafea93976c8c63a20b1ad147d9ad279d58fc639c9ba1c5fe" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.258306 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.444946 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5"] Nov 22 08:55:10 crc kubenswrapper[4735]: E1122 08:55:10.445598 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691ac120-c31b-4133-9847-4bb191808eb8" containerName="extract-utilities" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.445621 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="691ac120-c31b-4133-9847-4bb191808eb8" containerName="extract-utilities" Nov 22 08:55:10 crc kubenswrapper[4735]: E1122 08:55:10.445641 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691ac120-c31b-4133-9847-4bb191808eb8" containerName="registry-server" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.445650 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="691ac120-c31b-4133-9847-4bb191808eb8" containerName="registry-server" Nov 22 08:55:10 crc kubenswrapper[4735]: E1122 08:55:10.445680 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="registry-server" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.445689 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="registry-server" Nov 22 08:55:10 crc kubenswrapper[4735]: E1122 08:55:10.445710 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="extract-utilities" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.445720 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="extract-utilities" Nov 22 08:55:10 crc kubenswrapper[4735]: E1122 08:55:10.445734 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a393bb4a-7413-4a36-bc56-e6a18f17fffa" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.445742 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a393bb4a-7413-4a36-bc56-e6a18f17fffa" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 22 08:55:10 crc kubenswrapper[4735]: E1122 08:55:10.445760 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="extract-content" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.445766 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="extract-content" Nov 22 08:55:10 crc kubenswrapper[4735]: E1122 08:55:10.445779 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691ac120-c31b-4133-9847-4bb191808eb8" containerName="extract-content" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.445785 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="691ac120-c31b-4133-9847-4bb191808eb8" containerName="extract-content" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.446132 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="691ac120-c31b-4133-9847-4bb191808eb8" containerName="registry-server" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.446172 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a393bb4a-7413-4a36-bc56-e6a18f17fffa" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.446191 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="03f0db8b-0c08-4911-8b4b-0fbc571824aa" containerName="registry-server" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.447323 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.458510 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.458826 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.458982 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.459123 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.459277 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rd7h5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.462675 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5"] Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.523737 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t699s\" (UniqueName: \"kubernetes.io/projected/3a770eb5-beff-422b-ac8d-a98899d08a49-kube-api-access-t699s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.524016 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.524078 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.524111 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.524171 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.628117 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t699s\" (UniqueName: \"kubernetes.io/projected/3a770eb5-beff-422b-ac8d-a98899d08a49-kube-api-access-t699s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.628285 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.628518 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.628587 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.628741 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.794912 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.795190 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.795338 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t699s\" (UniqueName: \"kubernetes.io/projected/3a770eb5-beff-422b-ac8d-a98899d08a49-kube-api-access-t699s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.796407 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:10 crc kubenswrapper[4735]: I1122 08:55:10.806954 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pnpt5\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:11 crc kubenswrapper[4735]: I1122 08:55:11.101828 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:11 crc kubenswrapper[4735]: I1122 08:55:11.766205 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5"] Nov 22 08:55:12 crc kubenswrapper[4735]: I1122 08:55:12.290455 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" event={"ID":"3a770eb5-beff-422b-ac8d-a98899d08a49","Type":"ContainerStarted","Data":"20d0b21cc8d316a4167db338b9128b055d3fca243a134bed8a0e20dcfaf9bd47"} Nov 22 08:55:13 crc kubenswrapper[4735]: I1122 08:55:13.269410 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:55:13 crc kubenswrapper[4735]: E1122 08:55:13.271752 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:55:14 crc kubenswrapper[4735]: I1122 08:55:14.326448 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" event={"ID":"3a770eb5-beff-422b-ac8d-a98899d08a49","Type":"ContainerStarted","Data":"88512d458f64aea6c7a01823e7de62233ef3ab48a2f671afb5592b6fe1020d15"} Nov 22 08:55:14 crc kubenswrapper[4735]: I1122 08:55:14.352614 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" podStartSLOduration=2.946886713 podStartE2EDuration="4.352592385s" podCreationTimestamp="2025-11-22 08:55:10 +0000 UTC" firstStartedPulling="2025-11-22 08:55:11.787729604 +0000 UTC m=+3133.392068239" lastFinishedPulling="2025-11-22 08:55:13.193435276 +0000 UTC m=+3134.797773911" observedRunningTime="2025-11-22 08:55:14.342227081 +0000 UTC m=+3135.946565706" watchObservedRunningTime="2025-11-22 08:55:14.352592385 +0000 UTC m=+3135.956930990" Nov 22 08:55:28 crc kubenswrapper[4735]: I1122 08:55:28.264402 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:55:28 crc kubenswrapper[4735]: E1122 08:55:28.265231 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:55:30 crc kubenswrapper[4735]: I1122 08:55:30.497954 4735 generic.go:334] "Generic (PLEG): container finished" podID="3a770eb5-beff-422b-ac8d-a98899d08a49" containerID="88512d458f64aea6c7a01823e7de62233ef3ab48a2f671afb5592b6fe1020d15" exitCode=0 Nov 22 08:55:30 crc kubenswrapper[4735]: I1122 08:55:30.498291 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" event={"ID":"3a770eb5-beff-422b-ac8d-a98899d08a49","Type":"ContainerDied","Data":"88512d458f64aea6c7a01823e7de62233ef3ab48a2f671afb5592b6fe1020d15"} Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.016899 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.207377 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-ssh-key\") pod \"3a770eb5-beff-422b-ac8d-a98899d08a49\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.207433 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-1\") pod \"3a770eb5-beff-422b-ac8d-a98899d08a49\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.207517 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t699s\" (UniqueName: \"kubernetes.io/projected/3a770eb5-beff-422b-ac8d-a98899d08a49-kube-api-access-t699s\") pod \"3a770eb5-beff-422b-ac8d-a98899d08a49\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.207764 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-0\") pod \"3a770eb5-beff-422b-ac8d-a98899d08a49\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.207832 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-inventory\") pod \"3a770eb5-beff-422b-ac8d-a98899d08a49\" (UID: \"3a770eb5-beff-422b-ac8d-a98899d08a49\") " Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.213290 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a770eb5-beff-422b-ac8d-a98899d08a49-kube-api-access-t699s" (OuterVolumeSpecName: "kube-api-access-t699s") pod "3a770eb5-beff-422b-ac8d-a98899d08a49" (UID: "3a770eb5-beff-422b-ac8d-a98899d08a49"). InnerVolumeSpecName "kube-api-access-t699s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.245700 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "3a770eb5-beff-422b-ac8d-a98899d08a49" (UID: "3a770eb5-beff-422b-ac8d-a98899d08a49"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.250116 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-inventory" (OuterVolumeSpecName: "inventory") pod "3a770eb5-beff-422b-ac8d-a98899d08a49" (UID: "3a770eb5-beff-422b-ac8d-a98899d08a49"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.257244 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "3a770eb5-beff-422b-ac8d-a98899d08a49" (UID: "3a770eb5-beff-422b-ac8d-a98899d08a49"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.258922 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a770eb5-beff-422b-ac8d-a98899d08a49" (UID: "3a770eb5-beff-422b-ac8d-a98899d08a49"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.311116 4735 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.311156 4735 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.311166 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.311174 4735 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a770eb5-beff-422b-ac8d-a98899d08a49-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.311209 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t699s\" (UniqueName: \"kubernetes.io/projected/3a770eb5-beff-422b-ac8d-a98899d08a49-kube-api-access-t699s\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.525196 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" event={"ID":"3a770eb5-beff-422b-ac8d-a98899d08a49","Type":"ContainerDied","Data":"20d0b21cc8d316a4167db338b9128b055d3fca243a134bed8a0e20dcfaf9bd47"} Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.525541 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20d0b21cc8d316a4167db338b9128b055d3fca243a134bed8a0e20dcfaf9bd47" Nov 22 08:55:32 crc kubenswrapper[4735]: I1122 08:55:32.525616 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pnpt5" Nov 22 08:55:39 crc kubenswrapper[4735]: I1122 08:55:39.276393 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:55:39 crc kubenswrapper[4735]: E1122 08:55:39.278028 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:55:42 crc kubenswrapper[4735]: E1122 08:55:42.499127 4735 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.236s" Nov 22 08:55:51 crc kubenswrapper[4735]: I1122 08:55:51.267758 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:55:51 crc kubenswrapper[4735]: E1122 08:55:51.268800 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:55:51 crc kubenswrapper[4735]: I1122 08:55:51.758023 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="be212b2c-6753-40d9-9bee-0dfe2ff10dbc" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 22 08:55:52 crc kubenswrapper[4735]: I1122 08:55:52.758133 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="a6c75463-22e2-416d-a878-d2998e83c460" containerName="prometheus" probeResult="failure" output="command timed out" Nov 22 08:56:03 crc kubenswrapper[4735]: I1122 08:56:03.265148 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:56:03 crc kubenswrapper[4735]: E1122 08:56:03.266320 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:56:15 crc kubenswrapper[4735]: I1122 08:56:15.263780 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:56:15 crc kubenswrapper[4735]: E1122 08:56:15.264645 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:56:29 crc kubenswrapper[4735]: I1122 08:56:29.280273 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:56:29 crc kubenswrapper[4735]: E1122 08:56:29.281847 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:56:42 crc kubenswrapper[4735]: I1122 08:56:42.263590 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:56:42 crc kubenswrapper[4735]: E1122 08:56:42.264360 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 08:56:54 crc kubenswrapper[4735]: I1122 08:56:54.263842 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 08:56:54 crc kubenswrapper[4735]: I1122 08:56:54.592614 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"34d2765386b26e41396a345fe0d01f7e6d0e3ea688c98022a29c57b8828fb396"} Nov 22 08:57:42 crc kubenswrapper[4735]: I1122 08:57:42.712640 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-d55fd4b4f-24s49" podUID="b3fc589b-feac-438c-b008-8911417ccfa7" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 22 08:59:16 crc kubenswrapper[4735]: I1122 08:59:16.131219 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:59:16 crc kubenswrapper[4735]: I1122 08:59:16.131726 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:59:46 crc kubenswrapper[4735]: I1122 08:59:46.132029 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:59:46 crc kubenswrapper[4735]: I1122 08:59:46.132450 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.181818 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p"] Nov 22 09:00:00 crc kubenswrapper[4735]: E1122 09:00:00.182974 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a770eb5-beff-422b-ac8d-a98899d08a49" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.182993 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a770eb5-beff-422b-ac8d-a98899d08a49" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.183330 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a770eb5-beff-422b-ac8d-a98899d08a49" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.184368 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.188134 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.188480 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.194575 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p"] Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.279018 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eaf0172f-b54d-493e-8e69-aaae5d05459c-secret-volume\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.279145 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eaf0172f-b54d-493e-8e69-aaae5d05459c-config-volume\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.279176 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vclp6\" (UniqueName: \"kubernetes.io/projected/eaf0172f-b54d-493e-8e69-aaae5d05459c-kube-api-access-vclp6\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.381020 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eaf0172f-b54d-493e-8e69-aaae5d05459c-secret-volume\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.381183 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eaf0172f-b54d-493e-8e69-aaae5d05459c-config-volume\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.381211 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vclp6\" (UniqueName: \"kubernetes.io/projected/eaf0172f-b54d-493e-8e69-aaae5d05459c-kube-api-access-vclp6\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.382131 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eaf0172f-b54d-493e-8e69-aaae5d05459c-config-volume\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.387190 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eaf0172f-b54d-493e-8e69-aaae5d05459c-secret-volume\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.450992 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vclp6\" (UniqueName: \"kubernetes.io/projected/eaf0172f-b54d-493e-8e69-aaae5d05459c-kube-api-access-vclp6\") pod \"collect-profiles-29396700-znr6p\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:00 crc kubenswrapper[4735]: I1122 09:00:00.508693 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:01 crc kubenswrapper[4735]: I1122 09:00:01.032992 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p"] Nov 22 09:00:01 crc kubenswrapper[4735]: I1122 09:00:01.187668 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" event={"ID":"eaf0172f-b54d-493e-8e69-aaae5d05459c","Type":"ContainerStarted","Data":"151bdc548b5ac097fee2e15610a782a304a6b1118f59dd495aa91c99ea06fb18"} Nov 22 09:00:02 crc kubenswrapper[4735]: I1122 09:00:02.198942 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" event={"ID":"eaf0172f-b54d-493e-8e69-aaae5d05459c","Type":"ContainerStarted","Data":"e1199b3e5316f5bd1beed2d3ca40bce6c0a7d52ad048a1a24c9b7d9266efad54"} Nov 22 09:00:02 crc kubenswrapper[4735]: I1122 09:00:02.236330 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" podStartSLOduration=2.236310257 podStartE2EDuration="2.236310257s" podCreationTimestamp="2025-11-22 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 09:00:02.227595415 +0000 UTC m=+3423.831934020" watchObservedRunningTime="2025-11-22 09:00:02.236310257 +0000 UTC m=+3423.840648862" Nov 22 09:00:03 crc kubenswrapper[4735]: I1122 09:00:03.212057 4735 generic.go:334] "Generic (PLEG): container finished" podID="eaf0172f-b54d-493e-8e69-aaae5d05459c" containerID="e1199b3e5316f5bd1beed2d3ca40bce6c0a7d52ad048a1a24c9b7d9266efad54" exitCode=0 Nov 22 09:00:03 crc kubenswrapper[4735]: I1122 09:00:03.212136 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" event={"ID":"eaf0172f-b54d-493e-8e69-aaae5d05459c","Type":"ContainerDied","Data":"e1199b3e5316f5bd1beed2d3ca40bce6c0a7d52ad048a1a24c9b7d9266efad54"} Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.657634 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.794641 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eaf0172f-b54d-493e-8e69-aaae5d05459c-config-volume\") pod \"eaf0172f-b54d-493e-8e69-aaae5d05459c\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.794765 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eaf0172f-b54d-493e-8e69-aaae5d05459c-secret-volume\") pod \"eaf0172f-b54d-493e-8e69-aaae5d05459c\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.794928 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vclp6\" (UniqueName: \"kubernetes.io/projected/eaf0172f-b54d-493e-8e69-aaae5d05459c-kube-api-access-vclp6\") pod \"eaf0172f-b54d-493e-8e69-aaae5d05459c\" (UID: \"eaf0172f-b54d-493e-8e69-aaae5d05459c\") " Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.795662 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaf0172f-b54d-493e-8e69-aaae5d05459c-config-volume" (OuterVolumeSpecName: "config-volume") pod "eaf0172f-b54d-493e-8e69-aaae5d05459c" (UID: "eaf0172f-b54d-493e-8e69-aaae5d05459c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.796266 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eaf0172f-b54d-493e-8e69-aaae5d05459c-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.801919 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf0172f-b54d-493e-8e69-aaae5d05459c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eaf0172f-b54d-493e-8e69-aaae5d05459c" (UID: "eaf0172f-b54d-493e-8e69-aaae5d05459c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.803708 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf0172f-b54d-493e-8e69-aaae5d05459c-kube-api-access-vclp6" (OuterVolumeSpecName: "kube-api-access-vclp6") pod "eaf0172f-b54d-493e-8e69-aaae5d05459c" (UID: "eaf0172f-b54d-493e-8e69-aaae5d05459c"). InnerVolumeSpecName "kube-api-access-vclp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.898040 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eaf0172f-b54d-493e-8e69-aaae5d05459c-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:00:04 crc kubenswrapper[4735]: I1122 09:00:04.898077 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vclp6\" (UniqueName: \"kubernetes.io/projected/eaf0172f-b54d-493e-8e69-aaae5d05459c-kube-api-access-vclp6\") on node \"crc\" DevicePath \"\"" Nov 22 09:00:05 crc kubenswrapper[4735]: I1122 09:00:05.235128 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" event={"ID":"eaf0172f-b54d-493e-8e69-aaae5d05459c","Type":"ContainerDied","Data":"151bdc548b5ac097fee2e15610a782a304a6b1118f59dd495aa91c99ea06fb18"} Nov 22 09:00:05 crc kubenswrapper[4735]: I1122 09:00:05.235174 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="151bdc548b5ac097fee2e15610a782a304a6b1118f59dd495aa91c99ea06fb18" Nov 22 09:00:05 crc kubenswrapper[4735]: I1122 09:00:05.235237 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p" Nov 22 09:00:05 crc kubenswrapper[4735]: I1122 09:00:05.747655 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp"] Nov 22 09:00:05 crc kubenswrapper[4735]: I1122 09:00:05.757848 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-bxlgp"] Nov 22 09:00:07 crc kubenswrapper[4735]: I1122 09:00:07.430404 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d7eea70-9a64-4989-87a3-b4d6059cb49a" path="/var/lib/kubelet/pods/6d7eea70-9a64-4989-87a3-b4d6059cb49a/volumes" Nov 22 09:00:09 crc kubenswrapper[4735]: I1122 09:00:09.391201 4735 scope.go:117] "RemoveContainer" containerID="02294add5e71d912d138c2807debf2933a4179cd8a1179a36e6c5dd982f94773" Nov 22 09:00:16 crc kubenswrapper[4735]: I1122 09:00:16.131057 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:00:16 crc kubenswrapper[4735]: I1122 09:00:16.131612 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:00:16 crc kubenswrapper[4735]: I1122 09:00:16.131658 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:00:16 crc kubenswrapper[4735]: I1122 09:00:16.132629 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34d2765386b26e41396a345fe0d01f7e6d0e3ea688c98022a29c57b8828fb396"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:00:16 crc kubenswrapper[4735]: I1122 09:00:16.132676 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://34d2765386b26e41396a345fe0d01f7e6d0e3ea688c98022a29c57b8828fb396" gracePeriod=600 Nov 22 09:00:16 crc kubenswrapper[4735]: I1122 09:00:16.361975 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="34d2765386b26e41396a345fe0d01f7e6d0e3ea688c98022a29c57b8828fb396" exitCode=0 Nov 22 09:00:16 crc kubenswrapper[4735]: I1122 09:00:16.362341 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"34d2765386b26e41396a345fe0d01f7e6d0e3ea688c98022a29c57b8828fb396"} Nov 22 09:00:16 crc kubenswrapper[4735]: I1122 09:00:16.362383 4735 scope.go:117] "RemoveContainer" containerID="13958669d668c24eb9de72b1711964eba4ad2a374a47a99348b0b3886abf01e2" Nov 22 09:00:16 crc kubenswrapper[4735]: E1122 09:00:16.396665 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7945a41_afa2_43b7_95a2_085e3b62e64b.slice/crio-34d2765386b26e41396a345fe0d01f7e6d0e3ea688c98022a29c57b8828fb396.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7945a41_afa2_43b7_95a2_085e3b62e64b.slice/crio-conmon-34d2765386b26e41396a345fe0d01f7e6d0e3ea688c98022a29c57b8828fb396.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:00:17 crc kubenswrapper[4735]: I1122 09:00:17.375309 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6"} Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.161194 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29396701-zh6s9"] Nov 22 09:01:00 crc kubenswrapper[4735]: E1122 09:01:00.162568 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf0172f-b54d-493e-8e69-aaae5d05459c" containerName="collect-profiles" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.162587 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf0172f-b54d-493e-8e69-aaae5d05459c" containerName="collect-profiles" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.163054 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf0172f-b54d-493e-8e69-aaae5d05459c" containerName="collect-profiles" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.165438 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.181957 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29396701-zh6s9"] Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.250900 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-config-data\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.250984 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8ml2\" (UniqueName: \"kubernetes.io/projected/0d2bda54-d543-40c3-8cfd-fb9e03080130-kube-api-access-d8ml2\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.251023 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-fernet-keys\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.251129 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-combined-ca-bundle\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.353860 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8ml2\" (UniqueName: \"kubernetes.io/projected/0d2bda54-d543-40c3-8cfd-fb9e03080130-kube-api-access-d8ml2\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.353933 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-fernet-keys\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.354069 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-combined-ca-bundle\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.354431 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-config-data\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.363274 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-fernet-keys\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.364046 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-config-data\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.372664 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-combined-ca-bundle\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.374584 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8ml2\" (UniqueName: \"kubernetes.io/projected/0d2bda54-d543-40c3-8cfd-fb9e03080130-kube-api-access-d8ml2\") pod \"keystone-cron-29396701-zh6s9\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.498495 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:00 crc kubenswrapper[4735]: I1122 09:01:00.974182 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29396701-zh6s9"] Nov 22 09:01:00 crc kubenswrapper[4735]: W1122 09:01:00.979144 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d2bda54_d543_40c3_8cfd_fb9e03080130.slice/crio-403e1738de468e65ccf6ee420cc331023937be61a2975100f9789701a92b232f WatchSource:0}: Error finding container 403e1738de468e65ccf6ee420cc331023937be61a2975100f9789701a92b232f: Status 404 returned error can't find the container with id 403e1738de468e65ccf6ee420cc331023937be61a2975100f9789701a92b232f Nov 22 09:01:01 crc kubenswrapper[4735]: I1122 09:01:01.885618 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396701-zh6s9" event={"ID":"0d2bda54-d543-40c3-8cfd-fb9e03080130","Type":"ContainerStarted","Data":"8edc1cecfdf8c82717897579bb52af8fccb3a99c89f2a03f701dd5463c9d4efc"} Nov 22 09:01:01 crc kubenswrapper[4735]: I1122 09:01:01.886002 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396701-zh6s9" event={"ID":"0d2bda54-d543-40c3-8cfd-fb9e03080130","Type":"ContainerStarted","Data":"403e1738de468e65ccf6ee420cc331023937be61a2975100f9789701a92b232f"} Nov 22 09:01:01 crc kubenswrapper[4735]: I1122 09:01:01.904704 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29396701-zh6s9" podStartSLOduration=1.904683248 podStartE2EDuration="1.904683248s" podCreationTimestamp="2025-11-22 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 09:01:01.901124807 +0000 UTC m=+3483.505463412" watchObservedRunningTime="2025-11-22 09:01:01.904683248 +0000 UTC m=+3483.509021853" Nov 22 09:01:04 crc kubenswrapper[4735]: I1122 09:01:04.918542 4735 generic.go:334] "Generic (PLEG): container finished" podID="0d2bda54-d543-40c3-8cfd-fb9e03080130" containerID="8edc1cecfdf8c82717897579bb52af8fccb3a99c89f2a03f701dd5463c9d4efc" exitCode=0 Nov 22 09:01:04 crc kubenswrapper[4735]: I1122 09:01:04.918585 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396701-zh6s9" event={"ID":"0d2bda54-d543-40c3-8cfd-fb9e03080130","Type":"ContainerDied","Data":"8edc1cecfdf8c82717897579bb52af8fccb3a99c89f2a03f701dd5463c9d4efc"} Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.371780 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.498396 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-fernet-keys\") pod \"0d2bda54-d543-40c3-8cfd-fb9e03080130\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.498513 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-combined-ca-bundle\") pod \"0d2bda54-d543-40c3-8cfd-fb9e03080130\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.498724 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-config-data\") pod \"0d2bda54-d543-40c3-8cfd-fb9e03080130\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.498795 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8ml2\" (UniqueName: \"kubernetes.io/projected/0d2bda54-d543-40c3-8cfd-fb9e03080130-kube-api-access-d8ml2\") pod \"0d2bda54-d543-40c3-8cfd-fb9e03080130\" (UID: \"0d2bda54-d543-40c3-8cfd-fb9e03080130\") " Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.504941 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0d2bda54-d543-40c3-8cfd-fb9e03080130" (UID: "0d2bda54-d543-40c3-8cfd-fb9e03080130"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.505574 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d2bda54-d543-40c3-8cfd-fb9e03080130-kube-api-access-d8ml2" (OuterVolumeSpecName: "kube-api-access-d8ml2") pod "0d2bda54-d543-40c3-8cfd-fb9e03080130" (UID: "0d2bda54-d543-40c3-8cfd-fb9e03080130"). InnerVolumeSpecName "kube-api-access-d8ml2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.537964 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d2bda54-d543-40c3-8cfd-fb9e03080130" (UID: "0d2bda54-d543-40c3-8cfd-fb9e03080130"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.583139 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-config-data" (OuterVolumeSpecName: "config-data") pod "0d2bda54-d543-40c3-8cfd-fb9e03080130" (UID: "0d2bda54-d543-40c3-8cfd-fb9e03080130"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.602679 4735 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.602729 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.602748 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8ml2\" (UniqueName: \"kubernetes.io/projected/0d2bda54-d543-40c3-8cfd-fb9e03080130-kube-api-access-d8ml2\") on node \"crc\" DevicePath \"\"" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.602770 4735 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0d2bda54-d543-40c3-8cfd-fb9e03080130-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.956871 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396701-zh6s9" event={"ID":"0d2bda54-d543-40c3-8cfd-fb9e03080130","Type":"ContainerDied","Data":"403e1738de468e65ccf6ee420cc331023937be61a2975100f9789701a92b232f"} Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.957270 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="403e1738de468e65ccf6ee420cc331023937be61a2975100f9789701a92b232f" Nov 22 09:01:06 crc kubenswrapper[4735]: I1122 09:01:06.957360 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396701-zh6s9" Nov 22 09:02:16 crc kubenswrapper[4735]: I1122 09:02:16.134054 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:02:16 crc kubenswrapper[4735]: I1122 09:02:16.134709 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:02:46 crc kubenswrapper[4735]: I1122 09:02:46.131068 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:02:46 crc kubenswrapper[4735]: I1122 09:02:46.131606 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.654166 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c2tws"] Nov 22 09:03:03 crc kubenswrapper[4735]: E1122 09:03:03.655351 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2bda54-d543-40c3-8cfd-fb9e03080130" containerName="keystone-cron" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.655367 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2bda54-d543-40c3-8cfd-fb9e03080130" containerName="keystone-cron" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.655850 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d2bda54-d543-40c3-8cfd-fb9e03080130" containerName="keystone-cron" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.667983 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.672422 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c2tws"] Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.823928 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-catalog-content\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.824095 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-utilities\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.824435 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbgn6\" (UniqueName: \"kubernetes.io/projected/c356c5f4-63af-44ac-93d5-fc765a5144a3-kube-api-access-jbgn6\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.926424 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-catalog-content\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.926849 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-utilities\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.927002 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbgn6\" (UniqueName: \"kubernetes.io/projected/c356c5f4-63af-44ac-93d5-fc765a5144a3-kube-api-access-jbgn6\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.927753 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-utilities\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.927906 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-catalog-content\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.973509 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbgn6\" (UniqueName: \"kubernetes.io/projected/c356c5f4-63af-44ac-93d5-fc765a5144a3-kube-api-access-jbgn6\") pod \"redhat-operators-c2tws\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:03 crc kubenswrapper[4735]: I1122 09:03:03.999880 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:04 crc kubenswrapper[4735]: I1122 09:03:04.622249 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c2tws"] Nov 22 09:03:05 crc kubenswrapper[4735]: I1122 09:03:05.361538 4735 generic.go:334] "Generic (PLEG): container finished" podID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerID="218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c" exitCode=0 Nov 22 09:03:05 crc kubenswrapper[4735]: I1122 09:03:05.361601 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2tws" event={"ID":"c356c5f4-63af-44ac-93d5-fc765a5144a3","Type":"ContainerDied","Data":"218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c"} Nov 22 09:03:05 crc kubenswrapper[4735]: I1122 09:03:05.362056 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2tws" event={"ID":"c356c5f4-63af-44ac-93d5-fc765a5144a3","Type":"ContainerStarted","Data":"5b04ae06e70d6b77fd676936357f73a98cbe248b121b1810e0cbab9c633b5026"} Nov 22 09:03:05 crc kubenswrapper[4735]: I1122 09:03:05.363949 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:03:06 crc kubenswrapper[4735]: I1122 09:03:06.373701 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2tws" event={"ID":"c356c5f4-63af-44ac-93d5-fc765a5144a3","Type":"ContainerStarted","Data":"18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db"} Nov 22 09:03:13 crc kubenswrapper[4735]: I1122 09:03:13.453439 4735 generic.go:334] "Generic (PLEG): container finished" podID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerID="18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db" exitCode=0 Nov 22 09:03:13 crc kubenswrapper[4735]: I1122 09:03:13.453491 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2tws" event={"ID":"c356c5f4-63af-44ac-93d5-fc765a5144a3","Type":"ContainerDied","Data":"18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db"} Nov 22 09:03:14 crc kubenswrapper[4735]: I1122 09:03:14.468253 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2tws" event={"ID":"c356c5f4-63af-44ac-93d5-fc765a5144a3","Type":"ContainerStarted","Data":"9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501"} Nov 22 09:03:14 crc kubenswrapper[4735]: I1122 09:03:14.491413 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c2tws" podStartSLOduration=2.756769625 podStartE2EDuration="11.491390158s" podCreationTimestamp="2025-11-22 09:03:03 +0000 UTC" firstStartedPulling="2025-11-22 09:03:05.363675829 +0000 UTC m=+3606.968014434" lastFinishedPulling="2025-11-22 09:03:14.098296362 +0000 UTC m=+3615.702634967" observedRunningTime="2025-11-22 09:03:14.485409466 +0000 UTC m=+3616.089748071" watchObservedRunningTime="2025-11-22 09:03:14.491390158 +0000 UTC m=+3616.095728773" Nov 22 09:03:16 crc kubenswrapper[4735]: I1122 09:03:16.132128 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:03:16 crc kubenswrapper[4735]: I1122 09:03:16.132370 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:03:16 crc kubenswrapper[4735]: I1122 09:03:16.132416 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:03:16 crc kubenswrapper[4735]: I1122 09:03:16.133345 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:03:16 crc kubenswrapper[4735]: I1122 09:03:16.133406 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" gracePeriod=600 Nov 22 09:03:16 crc kubenswrapper[4735]: E1122 09:03:16.757340 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:03:17 crc kubenswrapper[4735]: I1122 09:03:17.503784 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" exitCode=0 Nov 22 09:03:17 crc kubenswrapper[4735]: I1122 09:03:17.503854 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6"} Nov 22 09:03:17 crc kubenswrapper[4735]: I1122 09:03:17.503899 4735 scope.go:117] "RemoveContainer" containerID="34d2765386b26e41396a345fe0d01f7e6d0e3ea688c98022a29c57b8828fb396" Nov 22 09:03:17 crc kubenswrapper[4735]: I1122 09:03:17.505023 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:03:17 crc kubenswrapper[4735]: E1122 09:03:17.505593 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:03:24 crc kubenswrapper[4735]: I1122 09:03:24.000953 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:24 crc kubenswrapper[4735]: I1122 09:03:24.003212 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:25 crc kubenswrapper[4735]: I1122 09:03:25.065521 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c2tws" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="registry-server" probeResult="failure" output=< Nov 22 09:03:25 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:03:25 crc kubenswrapper[4735]: > Nov 22 09:03:30 crc kubenswrapper[4735]: I1122 09:03:30.264029 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:03:30 crc kubenswrapper[4735]: E1122 09:03:30.265153 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:03:34 crc kubenswrapper[4735]: I1122 09:03:34.055202 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:34 crc kubenswrapper[4735]: I1122 09:03:34.126502 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:34 crc kubenswrapper[4735]: I1122 09:03:34.842189 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c2tws"] Nov 22 09:03:35 crc kubenswrapper[4735]: I1122 09:03:35.703720 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c2tws" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="registry-server" containerID="cri-o://9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501" gracePeriod=2 Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.286665 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.316519 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-catalog-content\") pod \"c356c5f4-63af-44ac-93d5-fc765a5144a3\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.316570 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-utilities\") pod \"c356c5f4-63af-44ac-93d5-fc765a5144a3\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.316607 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbgn6\" (UniqueName: \"kubernetes.io/projected/c356c5f4-63af-44ac-93d5-fc765a5144a3-kube-api-access-jbgn6\") pod \"c356c5f4-63af-44ac-93d5-fc765a5144a3\" (UID: \"c356c5f4-63af-44ac-93d5-fc765a5144a3\") " Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.320995 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-utilities" (OuterVolumeSpecName: "utilities") pod "c356c5f4-63af-44ac-93d5-fc765a5144a3" (UID: "c356c5f4-63af-44ac-93d5-fc765a5144a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.325362 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c356c5f4-63af-44ac-93d5-fc765a5144a3-kube-api-access-jbgn6" (OuterVolumeSpecName: "kube-api-access-jbgn6") pod "c356c5f4-63af-44ac-93d5-fc765a5144a3" (UID: "c356c5f4-63af-44ac-93d5-fc765a5144a3"). InnerVolumeSpecName "kube-api-access-jbgn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.417214 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c356c5f4-63af-44ac-93d5-fc765a5144a3" (UID: "c356c5f4-63af-44ac-93d5-fc765a5144a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.420574 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.420600 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c356c5f4-63af-44ac-93d5-fc765a5144a3-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.420610 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbgn6\" (UniqueName: \"kubernetes.io/projected/c356c5f4-63af-44ac-93d5-fc765a5144a3-kube-api-access-jbgn6\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.719087 4735 generic.go:334] "Generic (PLEG): container finished" podID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerID="9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501" exitCode=0 Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.719145 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2tws" event={"ID":"c356c5f4-63af-44ac-93d5-fc765a5144a3","Type":"ContainerDied","Data":"9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501"} Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.719435 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2tws" event={"ID":"c356c5f4-63af-44ac-93d5-fc765a5144a3","Type":"ContainerDied","Data":"5b04ae06e70d6b77fd676936357f73a98cbe248b121b1810e0cbab9c633b5026"} Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.719476 4735 scope.go:117] "RemoveContainer" containerID="9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.719178 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2tws" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.751563 4735 scope.go:117] "RemoveContainer" containerID="18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.761919 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c2tws"] Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.772504 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c2tws"] Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.783673 4735 scope.go:117] "RemoveContainer" containerID="218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.830440 4735 scope.go:117] "RemoveContainer" containerID="9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501" Nov 22 09:03:36 crc kubenswrapper[4735]: E1122 09:03:36.831083 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501\": container with ID starting with 9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501 not found: ID does not exist" containerID="9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.831115 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501"} err="failed to get container status \"9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501\": rpc error: code = NotFound desc = could not find container \"9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501\": container with ID starting with 9866ea26d29dfd97ed6f9a535e9c219d09f0907f7fbd33e05eab5c96684a6501 not found: ID does not exist" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.831144 4735 scope.go:117] "RemoveContainer" containerID="18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db" Nov 22 09:03:36 crc kubenswrapper[4735]: E1122 09:03:36.831397 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db\": container with ID starting with 18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db not found: ID does not exist" containerID="18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.831431 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db"} err="failed to get container status \"18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db\": rpc error: code = NotFound desc = could not find container \"18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db\": container with ID starting with 18e90eb40c1815ef6cdba03d182b19da14c992ffb8d9c8b3bb2008751764e4db not found: ID does not exist" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.831451 4735 scope.go:117] "RemoveContainer" containerID="218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c" Nov 22 09:03:36 crc kubenswrapper[4735]: E1122 09:03:36.832008 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c\": container with ID starting with 218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c not found: ID does not exist" containerID="218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c" Nov 22 09:03:36 crc kubenswrapper[4735]: I1122 09:03:36.832038 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c"} err="failed to get container status \"218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c\": rpc error: code = NotFound desc = could not find container \"218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c\": container with ID starting with 218c4fe02ce07145af92f84ec77c64f825d7efe2d4d13bd0cb4366a05e2bf34c not found: ID does not exist" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.053637 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mbrzp"] Nov 22 09:03:37 crc kubenswrapper[4735]: E1122 09:03:37.054268 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="extract-utilities" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.054291 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="extract-utilities" Nov 22 09:03:37 crc kubenswrapper[4735]: E1122 09:03:37.054347 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="registry-server" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.054356 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="registry-server" Nov 22 09:03:37 crc kubenswrapper[4735]: E1122 09:03:37.054376 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="extract-content" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.054386 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="extract-content" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.054690 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" containerName="registry-server" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.065911 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbrzp"] Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.072751 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.137378 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfp4w\" (UniqueName: \"kubernetes.io/projected/31f811ba-fac7-4eac-863a-dec2bc39750f-kube-api-access-zfp4w\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.137554 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-catalog-content\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.137596 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-utilities\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.239687 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-catalog-content\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.239755 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-utilities\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.239905 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfp4w\" (UniqueName: \"kubernetes.io/projected/31f811ba-fac7-4eac-863a-dec2bc39750f-kube-api-access-zfp4w\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.240663 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-catalog-content\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.240761 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-utilities\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.261235 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfp4w\" (UniqueName: \"kubernetes.io/projected/31f811ba-fac7-4eac-863a-dec2bc39750f-kube-api-access-zfp4w\") pod \"certified-operators-mbrzp\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.278827 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c356c5f4-63af-44ac-93d5-fc765a5144a3" path="/var/lib/kubelet/pods/c356c5f4-63af-44ac-93d5-fc765a5144a3/volumes" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.403952 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:37 crc kubenswrapper[4735]: I1122 09:03:37.956836 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbrzp"] Nov 22 09:03:38 crc kubenswrapper[4735]: I1122 09:03:38.758775 4735 generic.go:334] "Generic (PLEG): container finished" podID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerID="5678181fa585b520211c6a5ca0a6899a45cc70e8137989b9f5d02da6d7dce9b2" exitCode=0 Nov 22 09:03:38 crc kubenswrapper[4735]: I1122 09:03:38.758851 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbrzp" event={"ID":"31f811ba-fac7-4eac-863a-dec2bc39750f","Type":"ContainerDied","Data":"5678181fa585b520211c6a5ca0a6899a45cc70e8137989b9f5d02da6d7dce9b2"} Nov 22 09:03:38 crc kubenswrapper[4735]: I1122 09:03:38.759082 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbrzp" event={"ID":"31f811ba-fac7-4eac-863a-dec2bc39750f","Type":"ContainerStarted","Data":"5aa9dac8b60d32f51025f89d9739856045e6628015708e43c3c7ced0eba3ec7c"} Nov 22 09:03:39 crc kubenswrapper[4735]: I1122 09:03:39.774040 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbrzp" event={"ID":"31f811ba-fac7-4eac-863a-dec2bc39750f","Type":"ContainerStarted","Data":"cc85b48ec4b3c485aa284c4685c97c86f369c09cb8d42e19c313246aee107432"} Nov 22 09:03:41 crc kubenswrapper[4735]: I1122 09:03:41.810643 4735 generic.go:334] "Generic (PLEG): container finished" podID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerID="cc85b48ec4b3c485aa284c4685c97c86f369c09cb8d42e19c313246aee107432" exitCode=0 Nov 22 09:03:41 crc kubenswrapper[4735]: I1122 09:03:41.810724 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbrzp" event={"ID":"31f811ba-fac7-4eac-863a-dec2bc39750f","Type":"ContainerDied","Data":"cc85b48ec4b3c485aa284c4685c97c86f369c09cb8d42e19c313246aee107432"} Nov 22 09:03:42 crc kubenswrapper[4735]: I1122 09:03:42.830132 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbrzp" event={"ID":"31f811ba-fac7-4eac-863a-dec2bc39750f","Type":"ContainerStarted","Data":"ee00dc7fcb08a634727a25017859e5b13cbcb2436fc8919c716838c5adfdf77c"} Nov 22 09:03:42 crc kubenswrapper[4735]: I1122 09:03:42.856230 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mbrzp" podStartSLOduration=2.417910339 podStartE2EDuration="5.85620467s" podCreationTimestamp="2025-11-22 09:03:37 +0000 UTC" firstStartedPulling="2025-11-22 09:03:38.761444215 +0000 UTC m=+3640.365782820" lastFinishedPulling="2025-11-22 09:03:42.199738546 +0000 UTC m=+3643.804077151" observedRunningTime="2025-11-22 09:03:42.844431802 +0000 UTC m=+3644.448770417" watchObservedRunningTime="2025-11-22 09:03:42.85620467 +0000 UTC m=+3644.460543285" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.702678 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2f4k8"] Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.705889 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.719404 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2f4k8"] Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.835772 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-utilities\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.836168 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-catalog-content\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.836276 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz965\" (UniqueName: \"kubernetes.io/projected/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-kube-api-access-bz965\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.938494 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-catalog-content\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.938575 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz965\" (UniqueName: \"kubernetes.io/projected/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-kube-api-access-bz965\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.938795 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-utilities\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.938965 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-catalog-content\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.939256 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-utilities\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:44 crc kubenswrapper[4735]: I1122 09:03:44.961699 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz965\" (UniqueName: \"kubernetes.io/projected/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-kube-api-access-bz965\") pod \"community-operators-2f4k8\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:45 crc kubenswrapper[4735]: I1122 09:03:45.039842 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:45 crc kubenswrapper[4735]: I1122 09:03:45.264947 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:03:45 crc kubenswrapper[4735]: E1122 09:03:45.266113 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:03:45 crc kubenswrapper[4735]: I1122 09:03:45.609727 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2f4k8"] Nov 22 09:03:45 crc kubenswrapper[4735]: I1122 09:03:45.860728 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f4k8" event={"ID":"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6","Type":"ContainerStarted","Data":"b606074be2417079357e914a9250995f575edca86f8e8cc6340fa49336b6afc8"} Nov 22 09:03:46 crc kubenswrapper[4735]: I1122 09:03:46.875184 4735 generic.go:334] "Generic (PLEG): container finished" podID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerID="955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c" exitCode=0 Nov 22 09:03:46 crc kubenswrapper[4735]: I1122 09:03:46.875230 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f4k8" event={"ID":"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6","Type":"ContainerDied","Data":"955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c"} Nov 22 09:03:47 crc kubenswrapper[4735]: I1122 09:03:47.405418 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:47 crc kubenswrapper[4735]: I1122 09:03:47.405745 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:47 crc kubenswrapper[4735]: I1122 09:03:47.482872 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:47 crc kubenswrapper[4735]: I1122 09:03:47.891003 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f4k8" event={"ID":"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6","Type":"ContainerStarted","Data":"5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3"} Nov 22 09:03:47 crc kubenswrapper[4735]: I1122 09:03:47.943748 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:49 crc kubenswrapper[4735]: I1122 09:03:49.840619 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbrzp"] Nov 22 09:03:49 crc kubenswrapper[4735]: I1122 09:03:49.917928 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mbrzp" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerName="registry-server" containerID="cri-o://ee00dc7fcb08a634727a25017859e5b13cbcb2436fc8919c716838c5adfdf77c" gracePeriod=2 Nov 22 09:03:50 crc kubenswrapper[4735]: I1122 09:03:50.929142 4735 generic.go:334] "Generic (PLEG): container finished" podID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerID="ee00dc7fcb08a634727a25017859e5b13cbcb2436fc8919c716838c5adfdf77c" exitCode=0 Nov 22 09:03:50 crc kubenswrapper[4735]: I1122 09:03:50.929176 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbrzp" event={"ID":"31f811ba-fac7-4eac-863a-dec2bc39750f","Type":"ContainerDied","Data":"ee00dc7fcb08a634727a25017859e5b13cbcb2436fc8919c716838c5adfdf77c"} Nov 22 09:03:50 crc kubenswrapper[4735]: I1122 09:03:50.929743 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbrzp" event={"ID":"31f811ba-fac7-4eac-863a-dec2bc39750f","Type":"ContainerDied","Data":"5aa9dac8b60d32f51025f89d9739856045e6628015708e43c3c7ced0eba3ec7c"} Nov 22 09:03:50 crc kubenswrapper[4735]: I1122 09:03:50.929759 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5aa9dac8b60d32f51025f89d9739856045e6628015708e43c3c7ced0eba3ec7c" Nov 22 09:03:50 crc kubenswrapper[4735]: I1122 09:03:50.991388 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.093954 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfp4w\" (UniqueName: \"kubernetes.io/projected/31f811ba-fac7-4eac-863a-dec2bc39750f-kube-api-access-zfp4w\") pod \"31f811ba-fac7-4eac-863a-dec2bc39750f\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.094080 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-catalog-content\") pod \"31f811ba-fac7-4eac-863a-dec2bc39750f\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.094210 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-utilities\") pod \"31f811ba-fac7-4eac-863a-dec2bc39750f\" (UID: \"31f811ba-fac7-4eac-863a-dec2bc39750f\") " Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.094917 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-utilities" (OuterVolumeSpecName: "utilities") pod "31f811ba-fac7-4eac-863a-dec2bc39750f" (UID: "31f811ba-fac7-4eac-863a-dec2bc39750f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.099785 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31f811ba-fac7-4eac-863a-dec2bc39750f-kube-api-access-zfp4w" (OuterVolumeSpecName: "kube-api-access-zfp4w") pod "31f811ba-fac7-4eac-863a-dec2bc39750f" (UID: "31f811ba-fac7-4eac-863a-dec2bc39750f"). InnerVolumeSpecName "kube-api-access-zfp4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.130279 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31f811ba-fac7-4eac-863a-dec2bc39750f" (UID: "31f811ba-fac7-4eac-863a-dec2bc39750f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.197123 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.197154 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f811ba-fac7-4eac-863a-dec2bc39750f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.197164 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfp4w\" (UniqueName: \"kubernetes.io/projected/31f811ba-fac7-4eac-863a-dec2bc39750f-kube-api-access-zfp4w\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.946005 4735 generic.go:334] "Generic (PLEG): container finished" podID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerID="5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3" exitCode=0 Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.946116 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f4k8" event={"ID":"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6","Type":"ContainerDied","Data":"5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3"} Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.946349 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbrzp" Nov 22 09:03:51 crc kubenswrapper[4735]: I1122 09:03:51.996372 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbrzp"] Nov 22 09:03:52 crc kubenswrapper[4735]: I1122 09:03:52.005854 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mbrzp"] Nov 22 09:03:52 crc kubenswrapper[4735]: I1122 09:03:52.970022 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f4k8" event={"ID":"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6","Type":"ContainerStarted","Data":"e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53"} Nov 22 09:03:53 crc kubenswrapper[4735]: I1122 09:03:53.008103 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2f4k8" podStartSLOduration=3.535693159 podStartE2EDuration="9.008063762s" podCreationTimestamp="2025-11-22 09:03:44 +0000 UTC" firstStartedPulling="2025-11-22 09:03:46.87719155 +0000 UTC m=+3648.481530165" lastFinishedPulling="2025-11-22 09:03:52.349562163 +0000 UTC m=+3653.953900768" observedRunningTime="2025-11-22 09:03:52.996357786 +0000 UTC m=+3654.600696391" watchObservedRunningTime="2025-11-22 09:03:53.008063762 +0000 UTC m=+3654.612402367" Nov 22 09:03:53 crc kubenswrapper[4735]: I1122 09:03:53.277345 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" path="/var/lib/kubelet/pods/31f811ba-fac7-4eac-863a-dec2bc39750f/volumes" Nov 22 09:03:55 crc kubenswrapper[4735]: I1122 09:03:55.041152 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:55 crc kubenswrapper[4735]: I1122 09:03:55.041504 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:03:55 crc kubenswrapper[4735]: I1122 09:03:55.114360 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:04:00 crc kubenswrapper[4735]: I1122 09:04:00.263488 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:04:00 crc kubenswrapper[4735]: E1122 09:04:00.264182 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:04:05 crc kubenswrapper[4735]: I1122 09:04:05.102008 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:04:05 crc kubenswrapper[4735]: I1122 09:04:05.168720 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2f4k8"] Nov 22 09:04:06 crc kubenswrapper[4735]: I1122 09:04:06.121816 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2f4k8" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerName="registry-server" containerID="cri-o://e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53" gracePeriod=2 Nov 22 09:04:06 crc kubenswrapper[4735]: I1122 09:04:06.781889 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:04:06 crc kubenswrapper[4735]: I1122 09:04:06.972479 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-catalog-content\") pod \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " Nov 22 09:04:06 crc kubenswrapper[4735]: I1122 09:04:06.972542 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-utilities\") pod \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " Nov 22 09:04:06 crc kubenswrapper[4735]: I1122 09:04:06.972604 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz965\" (UniqueName: \"kubernetes.io/projected/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-kube-api-access-bz965\") pod \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\" (UID: \"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6\") " Nov 22 09:04:06 crc kubenswrapper[4735]: I1122 09:04:06.973914 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-utilities" (OuterVolumeSpecName: "utilities") pod "5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" (UID: "5296b0d9-ee7c-4068-a1ec-5219bb46b3a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:04:06 crc kubenswrapper[4735]: I1122 09:04:06.981121 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-kube-api-access-bz965" (OuterVolumeSpecName: "kube-api-access-bz965") pod "5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" (UID: "5296b0d9-ee7c-4068-a1ec-5219bb46b3a6"). InnerVolumeSpecName "kube-api-access-bz965". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.032411 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" (UID: "5296b0d9-ee7c-4068-a1ec-5219bb46b3a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.075852 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.075884 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.075894 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz965\" (UniqueName: \"kubernetes.io/projected/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6-kube-api-access-bz965\") on node \"crc\" DevicePath \"\"" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.134853 4735 generic.go:334] "Generic (PLEG): container finished" podID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerID="e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53" exitCode=0 Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.134896 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f4k8" event={"ID":"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6","Type":"ContainerDied","Data":"e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53"} Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.134925 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f4k8" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.134946 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f4k8" event={"ID":"5296b0d9-ee7c-4068-a1ec-5219bb46b3a6","Type":"ContainerDied","Data":"b606074be2417079357e914a9250995f575edca86f8e8cc6340fa49336b6afc8"} Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.134964 4735 scope.go:117] "RemoveContainer" containerID="e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.172197 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2f4k8"] Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.172277 4735 scope.go:117] "RemoveContainer" containerID="5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.182275 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2f4k8"] Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.207976 4735 scope.go:117] "RemoveContainer" containerID="955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.250978 4735 scope.go:117] "RemoveContainer" containerID="e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53" Nov 22 09:04:07 crc kubenswrapper[4735]: E1122 09:04:07.251605 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53\": container with ID starting with e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53 not found: ID does not exist" containerID="e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.251643 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53"} err="failed to get container status \"e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53\": rpc error: code = NotFound desc = could not find container \"e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53\": container with ID starting with e22b91f3050dbad14489a88d4d3c6146533c1088b7328b825545730c752a2d53 not found: ID does not exist" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.251668 4735 scope.go:117] "RemoveContainer" containerID="5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3" Nov 22 09:04:07 crc kubenswrapper[4735]: E1122 09:04:07.252337 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3\": container with ID starting with 5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3 not found: ID does not exist" containerID="5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.252376 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3"} err="failed to get container status \"5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3\": rpc error: code = NotFound desc = could not find container \"5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3\": container with ID starting with 5a2a0aaff9ccf3985c3602885c9dd47416c091870ad91d9f82c5aa61f8a894e3 not found: ID does not exist" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.252400 4735 scope.go:117] "RemoveContainer" containerID="955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c" Nov 22 09:04:07 crc kubenswrapper[4735]: E1122 09:04:07.252948 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c\": container with ID starting with 955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c not found: ID does not exist" containerID="955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.252976 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c"} err="failed to get container status \"955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c\": rpc error: code = NotFound desc = could not find container \"955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c\": container with ID starting with 955cc819e211b0d60c5b29f109431523c72ca090c8937e2d156f724830200d4c not found: ID does not exist" Nov 22 09:04:07 crc kubenswrapper[4735]: I1122 09:04:07.280033 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" path="/var/lib/kubelet/pods/5296b0d9-ee7c-4068-a1ec-5219bb46b3a6/volumes" Nov 22 09:04:14 crc kubenswrapper[4735]: I1122 09:04:14.263493 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:04:14 crc kubenswrapper[4735]: E1122 09:04:14.264398 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:04:29 crc kubenswrapper[4735]: I1122 09:04:29.272621 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:04:29 crc kubenswrapper[4735]: E1122 09:04:29.273557 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:04:40 crc kubenswrapper[4735]: I1122 09:04:40.263572 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:04:40 crc kubenswrapper[4735]: E1122 09:04:40.264351 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:04:55 crc kubenswrapper[4735]: I1122 09:04:55.263728 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:04:55 crc kubenswrapper[4735]: E1122 09:04:55.264531 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:05:10 crc kubenswrapper[4735]: I1122 09:05:10.263772 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:05:10 crc kubenswrapper[4735]: E1122 09:05:10.266071 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:05:21 crc kubenswrapper[4735]: I1122 09:05:21.263468 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:05:21 crc kubenswrapper[4735]: E1122 09:05:21.264121 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:05:35 crc kubenswrapper[4735]: I1122 09:05:35.372333 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 22 09:05:36 crc kubenswrapper[4735]: I1122 09:05:36.263591 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:05:36 crc kubenswrapper[4735]: E1122 09:05:36.263920 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:05:47 crc kubenswrapper[4735]: I1122 09:05:47.264859 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:05:47 crc kubenswrapper[4735]: E1122 09:05:47.266215 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:06:00 crc kubenswrapper[4735]: I1122 09:06:00.263596 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:06:00 crc kubenswrapper[4735]: E1122 09:06:00.265039 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:06:12 crc kubenswrapper[4735]: I1122 09:06:12.264012 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:06:12 crc kubenswrapper[4735]: E1122 09:06:12.264757 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:06:27 crc kubenswrapper[4735]: I1122 09:06:27.263380 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:06:27 crc kubenswrapper[4735]: E1122 09:06:27.264502 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:06:41 crc kubenswrapper[4735]: I1122 09:06:41.265626 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:06:41 crc kubenswrapper[4735]: E1122 09:06:41.266384 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:06:55 crc kubenswrapper[4735]: I1122 09:06:55.263665 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:06:55 crc kubenswrapper[4735]: E1122 09:06:55.264383 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:07:09 crc kubenswrapper[4735]: I1122 09:07:09.284547 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:07:09 crc kubenswrapper[4735]: E1122 09:07:09.285455 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:07:24 crc kubenswrapper[4735]: I1122 09:07:24.265795 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:07:24 crc kubenswrapper[4735]: E1122 09:07:24.267524 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:07:38 crc kubenswrapper[4735]: I1122 09:07:38.263412 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:07:38 crc kubenswrapper[4735]: E1122 09:07:38.264239 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:07:50 crc kubenswrapper[4735]: I1122 09:07:50.263548 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:07:50 crc kubenswrapper[4735]: E1122 09:07:50.264434 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:08:05 crc kubenswrapper[4735]: I1122 09:08:05.264369 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:08:05 crc kubenswrapper[4735]: E1122 09:08:05.265618 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:08:16 crc kubenswrapper[4735]: I1122 09:08:16.264335 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:08:16 crc kubenswrapper[4735]: E1122 09:08:16.265375 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:08:29 crc kubenswrapper[4735]: I1122 09:08:29.274960 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:08:30 crc kubenswrapper[4735]: I1122 09:08:30.225243 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"68aa55a00176dae0321a1f4209cd4c13a08313ec37415b0b62700c8b94230560"} Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.244013 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dw4lx"] Nov 22 09:09:06 crc kubenswrapper[4735]: E1122 09:09:06.245235 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerName="registry-server" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.245255 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerName="registry-server" Nov 22 09:09:06 crc kubenswrapper[4735]: E1122 09:09:06.245274 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerName="extract-content" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.245285 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerName="extract-content" Nov 22 09:09:06 crc kubenswrapper[4735]: E1122 09:09:06.245305 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerName="registry-server" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.245317 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerName="registry-server" Nov 22 09:09:06 crc kubenswrapper[4735]: E1122 09:09:06.245346 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerName="extract-content" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.245354 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerName="extract-content" Nov 22 09:09:06 crc kubenswrapper[4735]: E1122 09:09:06.245366 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerName="extract-utilities" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.245375 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerName="extract-utilities" Nov 22 09:09:06 crc kubenswrapper[4735]: E1122 09:09:06.245389 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerName="extract-utilities" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.245397 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerName="extract-utilities" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.245700 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5296b0d9-ee7c-4068-a1ec-5219bb46b3a6" containerName="registry-server" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.245726 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="31f811ba-fac7-4eac-863a-dec2bc39750f" containerName="registry-server" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.247944 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.292156 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dw4lx"] Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.362870 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-945mp\" (UniqueName: \"kubernetes.io/projected/885655fe-309b-487a-aec1-7d999c79729e-kube-api-access-945mp\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.362971 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-catalog-content\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.364125 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-utilities\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.466613 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-utilities\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.466827 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-945mp\" (UniqueName: \"kubernetes.io/projected/885655fe-309b-487a-aec1-7d999c79729e-kube-api-access-945mp\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.466881 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-catalog-content\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.467138 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-utilities\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.467634 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-catalog-content\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.517316 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-945mp\" (UniqueName: \"kubernetes.io/projected/885655fe-309b-487a-aec1-7d999c79729e-kube-api-access-945mp\") pod \"redhat-marketplace-dw4lx\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:06 crc kubenswrapper[4735]: I1122 09:09:06.576449 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:07 crc kubenswrapper[4735]: I1122 09:09:07.103313 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dw4lx"] Nov 22 09:09:07 crc kubenswrapper[4735]: I1122 09:09:07.647682 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dw4lx" event={"ID":"885655fe-309b-487a-aec1-7d999c79729e","Type":"ContainerStarted","Data":"c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e"} Nov 22 09:09:07 crc kubenswrapper[4735]: I1122 09:09:07.648038 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dw4lx" event={"ID":"885655fe-309b-487a-aec1-7d999c79729e","Type":"ContainerStarted","Data":"09ac79c1b0564dc61875a854d9fecbcff0d9681d37de19ffb4a6c3c833dfbda0"} Nov 22 09:09:08 crc kubenswrapper[4735]: I1122 09:09:08.662536 4735 generic.go:334] "Generic (PLEG): container finished" podID="885655fe-309b-487a-aec1-7d999c79729e" containerID="c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e" exitCode=0 Nov 22 09:09:08 crc kubenswrapper[4735]: I1122 09:09:08.662931 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dw4lx" event={"ID":"885655fe-309b-487a-aec1-7d999c79729e","Type":"ContainerDied","Data":"c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e"} Nov 22 09:09:08 crc kubenswrapper[4735]: I1122 09:09:08.665604 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:09:13 crc kubenswrapper[4735]: I1122 09:09:13.719162 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dw4lx" event={"ID":"885655fe-309b-487a-aec1-7d999c79729e","Type":"ContainerStarted","Data":"8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee"} Nov 22 09:09:14 crc kubenswrapper[4735]: I1122 09:09:14.731811 4735 generic.go:334] "Generic (PLEG): container finished" podID="885655fe-309b-487a-aec1-7d999c79729e" containerID="8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee" exitCode=0 Nov 22 09:09:14 crc kubenswrapper[4735]: I1122 09:09:14.731860 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dw4lx" event={"ID":"885655fe-309b-487a-aec1-7d999c79729e","Type":"ContainerDied","Data":"8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee"} Nov 22 09:09:17 crc kubenswrapper[4735]: I1122 09:09:17.794529 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dw4lx" event={"ID":"885655fe-309b-487a-aec1-7d999c79729e","Type":"ContainerStarted","Data":"f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d"} Nov 22 09:09:17 crc kubenswrapper[4735]: I1122 09:09:17.825787 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dw4lx" podStartSLOduration=3.598353556 podStartE2EDuration="11.825759092s" podCreationTimestamp="2025-11-22 09:09:06 +0000 UTC" firstStartedPulling="2025-11-22 09:09:08.665348052 +0000 UTC m=+3970.269686657" lastFinishedPulling="2025-11-22 09:09:16.892753588 +0000 UTC m=+3978.497092193" observedRunningTime="2025-11-22 09:09:17.81342294 +0000 UTC m=+3979.417761545" watchObservedRunningTime="2025-11-22 09:09:17.825759092 +0000 UTC m=+3979.430097717" Nov 22 09:09:26 crc kubenswrapper[4735]: I1122 09:09:26.579781 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:26 crc kubenswrapper[4735]: I1122 09:09:26.580564 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:26 crc kubenswrapper[4735]: I1122 09:09:26.633818 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:26 crc kubenswrapper[4735]: I1122 09:09:26.939525 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:26 crc kubenswrapper[4735]: I1122 09:09:26.999543 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dw4lx"] Nov 22 09:09:28 crc kubenswrapper[4735]: I1122 09:09:28.908125 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dw4lx" podUID="885655fe-309b-487a-aec1-7d999c79729e" containerName="registry-server" containerID="cri-o://f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d" gracePeriod=2 Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.458039 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.595496 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-utilities\") pod \"885655fe-309b-487a-aec1-7d999c79729e\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.595589 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-catalog-content\") pod \"885655fe-309b-487a-aec1-7d999c79729e\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.595621 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-945mp\" (UniqueName: \"kubernetes.io/projected/885655fe-309b-487a-aec1-7d999c79729e-kube-api-access-945mp\") pod \"885655fe-309b-487a-aec1-7d999c79729e\" (UID: \"885655fe-309b-487a-aec1-7d999c79729e\") " Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.596664 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-utilities" (OuterVolumeSpecName: "utilities") pod "885655fe-309b-487a-aec1-7d999c79729e" (UID: "885655fe-309b-487a-aec1-7d999c79729e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.603350 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/885655fe-309b-487a-aec1-7d999c79729e-kube-api-access-945mp" (OuterVolumeSpecName: "kube-api-access-945mp") pod "885655fe-309b-487a-aec1-7d999c79729e" (UID: "885655fe-309b-487a-aec1-7d999c79729e"). InnerVolumeSpecName "kube-api-access-945mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.617105 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "885655fe-309b-487a-aec1-7d999c79729e" (UID: "885655fe-309b-487a-aec1-7d999c79729e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.698626 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.698663 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/885655fe-309b-487a-aec1-7d999c79729e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.698678 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-945mp\" (UniqueName: \"kubernetes.io/projected/885655fe-309b-487a-aec1-7d999c79729e-kube-api-access-945mp\") on node \"crc\" DevicePath \"\"" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.923479 4735 generic.go:334] "Generic (PLEG): container finished" podID="885655fe-309b-487a-aec1-7d999c79729e" containerID="f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d" exitCode=0 Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.923525 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dw4lx" event={"ID":"885655fe-309b-487a-aec1-7d999c79729e","Type":"ContainerDied","Data":"f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d"} Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.923557 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dw4lx" event={"ID":"885655fe-309b-487a-aec1-7d999c79729e","Type":"ContainerDied","Data":"09ac79c1b0564dc61875a854d9fecbcff0d9681d37de19ffb4a6c3c833dfbda0"} Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.923577 4735 scope.go:117] "RemoveContainer" containerID="f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.923600 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dw4lx" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.950295 4735 scope.go:117] "RemoveContainer" containerID="8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee" Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.969060 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dw4lx"] Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.981680 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dw4lx"] Nov 22 09:09:29 crc kubenswrapper[4735]: I1122 09:09:29.988996 4735 scope.go:117] "RemoveContainer" containerID="c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e" Nov 22 09:09:30 crc kubenswrapper[4735]: I1122 09:09:30.052856 4735 scope.go:117] "RemoveContainer" containerID="f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d" Nov 22 09:09:30 crc kubenswrapper[4735]: E1122 09:09:30.053852 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d\": container with ID starting with f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d not found: ID does not exist" containerID="f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d" Nov 22 09:09:30 crc kubenswrapper[4735]: I1122 09:09:30.053892 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d"} err="failed to get container status \"f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d\": rpc error: code = NotFound desc = could not find container \"f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d\": container with ID starting with f6b5b4e8c558d74118608d279014780b540caea3de903a6e6f76c8e2dbb2540d not found: ID does not exist" Nov 22 09:09:30 crc kubenswrapper[4735]: I1122 09:09:30.053918 4735 scope.go:117] "RemoveContainer" containerID="8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee" Nov 22 09:09:30 crc kubenswrapper[4735]: E1122 09:09:30.054216 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee\": container with ID starting with 8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee not found: ID does not exist" containerID="8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee" Nov 22 09:09:30 crc kubenswrapper[4735]: I1122 09:09:30.054252 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee"} err="failed to get container status \"8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee\": rpc error: code = NotFound desc = could not find container \"8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee\": container with ID starting with 8eedcf4fcfb7b7eb5102acfedb5eae82b01266869940a5da9df914ed9ad694ee not found: ID does not exist" Nov 22 09:09:30 crc kubenswrapper[4735]: I1122 09:09:30.054276 4735 scope.go:117] "RemoveContainer" containerID="c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e" Nov 22 09:09:30 crc kubenswrapper[4735]: E1122 09:09:30.054525 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e\": container with ID starting with c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e not found: ID does not exist" containerID="c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e" Nov 22 09:09:30 crc kubenswrapper[4735]: I1122 09:09:30.054560 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e"} err="failed to get container status \"c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e\": rpc error: code = NotFound desc = could not find container \"c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e\": container with ID starting with c8c0aa0c87812d28eebf9a091f8ad929911b676f7e05ec88622d96504be7194e not found: ID does not exist" Nov 22 09:09:31 crc kubenswrapper[4735]: I1122 09:09:31.292338 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="885655fe-309b-487a-aec1-7d999c79729e" path="/var/lib/kubelet/pods/885655fe-309b-487a-aec1-7d999c79729e/volumes" Nov 22 09:10:09 crc kubenswrapper[4735]: I1122 09:10:09.834903 4735 scope.go:117] "RemoveContainer" containerID="ee00dc7fcb08a634727a25017859e5b13cbcb2436fc8919c716838c5adfdf77c" Nov 22 09:10:09 crc kubenswrapper[4735]: I1122 09:10:09.856297 4735 scope.go:117] "RemoveContainer" containerID="5678181fa585b520211c6a5ca0a6899a45cc70e8137989b9f5d02da6d7dce9b2" Nov 22 09:10:09 crc kubenswrapper[4735]: I1122 09:10:09.878794 4735 scope.go:117] "RemoveContainer" containerID="cc85b48ec4b3c485aa284c4685c97c86f369c09cb8d42e19c313246aee107432" Nov 22 09:10:46 crc kubenswrapper[4735]: I1122 09:10:46.131699 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:10:46 crc kubenswrapper[4735]: I1122 09:10:46.132395 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:11:16 crc kubenswrapper[4735]: I1122 09:11:16.132254 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:11:16 crc kubenswrapper[4735]: I1122 09:11:16.132828 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:11:46 crc kubenswrapper[4735]: I1122 09:11:46.132047 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:11:46 crc kubenswrapper[4735]: I1122 09:11:46.132650 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:11:46 crc kubenswrapper[4735]: I1122 09:11:46.132704 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:11:46 crc kubenswrapper[4735]: I1122 09:11:46.133656 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"68aa55a00176dae0321a1f4209cd4c13a08313ec37415b0b62700c8b94230560"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:11:46 crc kubenswrapper[4735]: I1122 09:11:46.133715 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://68aa55a00176dae0321a1f4209cd4c13a08313ec37415b0b62700c8b94230560" gracePeriod=600 Nov 22 09:11:46 crc kubenswrapper[4735]: I1122 09:11:46.453930 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="68aa55a00176dae0321a1f4209cd4c13a08313ec37415b0b62700c8b94230560" exitCode=0 Nov 22 09:11:46 crc kubenswrapper[4735]: I1122 09:11:46.453992 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"68aa55a00176dae0321a1f4209cd4c13a08313ec37415b0b62700c8b94230560"} Nov 22 09:11:46 crc kubenswrapper[4735]: I1122 09:11:46.454322 4735 scope.go:117] "RemoveContainer" containerID="96b5616f084b513eea1ea0d2e5543548f86eab9e7f2cec2c5f31e679d4a5b5f6" Nov 22 09:11:47 crc kubenswrapper[4735]: I1122 09:11:47.483302 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076"} Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.013735 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rxtdz"] Nov 22 09:13:18 crc kubenswrapper[4735]: E1122 09:13:18.015135 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="885655fe-309b-487a-aec1-7d999c79729e" containerName="extract-content" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.015156 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="885655fe-309b-487a-aec1-7d999c79729e" containerName="extract-content" Nov 22 09:13:18 crc kubenswrapper[4735]: E1122 09:13:18.015192 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="885655fe-309b-487a-aec1-7d999c79729e" containerName="registry-server" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.015209 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="885655fe-309b-487a-aec1-7d999c79729e" containerName="registry-server" Nov 22 09:13:18 crc kubenswrapper[4735]: E1122 09:13:18.015731 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="885655fe-309b-487a-aec1-7d999c79729e" containerName="extract-utilities" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.015751 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="885655fe-309b-487a-aec1-7d999c79729e" containerName="extract-utilities" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.016137 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="885655fe-309b-487a-aec1-7d999c79729e" containerName="registry-server" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.018518 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.029801 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rxtdz"] Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.123848 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-utilities\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.123942 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-catalog-content\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.123999 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnv8m\" (UniqueName: \"kubernetes.io/projected/07d59ecf-3c54-4f35-b13f-c52edfbb6091-kube-api-access-rnv8m\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.227055 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-utilities\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.227278 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-catalog-content\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.227367 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnv8m\" (UniqueName: \"kubernetes.io/projected/07d59ecf-3c54-4f35-b13f-c52edfbb6091-kube-api-access-rnv8m\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.227951 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-catalog-content\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.227963 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-utilities\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.247845 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnv8m\" (UniqueName: \"kubernetes.io/projected/07d59ecf-3c54-4f35-b13f-c52edfbb6091-kube-api-access-rnv8m\") pod \"redhat-operators-rxtdz\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.351182 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:18 crc kubenswrapper[4735]: I1122 09:13:18.878725 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rxtdz"] Nov 22 09:13:19 crc kubenswrapper[4735]: I1122 09:13:19.621956 4735 generic.go:334] "Generic (PLEG): container finished" podID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerID="27f7de275e90d12fdbdeacb62d87b0d37ed4fd839ad1c1dd5953dd7887684338" exitCode=0 Nov 22 09:13:19 crc kubenswrapper[4735]: I1122 09:13:19.622014 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxtdz" event={"ID":"07d59ecf-3c54-4f35-b13f-c52edfbb6091","Type":"ContainerDied","Data":"27f7de275e90d12fdbdeacb62d87b0d37ed4fd839ad1c1dd5953dd7887684338"} Nov 22 09:13:19 crc kubenswrapper[4735]: I1122 09:13:19.622271 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxtdz" event={"ID":"07d59ecf-3c54-4f35-b13f-c52edfbb6091","Type":"ContainerStarted","Data":"b058509c31032ab22595a78445a736d2a636234b66bf15d26a1a6636932d4964"} Nov 22 09:13:21 crc kubenswrapper[4735]: I1122 09:13:21.641569 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxtdz" event={"ID":"07d59ecf-3c54-4f35-b13f-c52edfbb6091","Type":"ContainerStarted","Data":"e1eef004b1c2d6c94c7362c66bc5afe2e951d13d214d58b3ed361c0a888a2e0f"} Nov 22 09:13:25 crc kubenswrapper[4735]: I1122 09:13:25.683907 4735 generic.go:334] "Generic (PLEG): container finished" podID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerID="e1eef004b1c2d6c94c7362c66bc5afe2e951d13d214d58b3ed361c0a888a2e0f" exitCode=0 Nov 22 09:13:25 crc kubenswrapper[4735]: I1122 09:13:25.684081 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxtdz" event={"ID":"07d59ecf-3c54-4f35-b13f-c52edfbb6091","Type":"ContainerDied","Data":"e1eef004b1c2d6c94c7362c66bc5afe2e951d13d214d58b3ed361c0a888a2e0f"} Nov 22 09:13:26 crc kubenswrapper[4735]: I1122 09:13:26.701658 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxtdz" event={"ID":"07d59ecf-3c54-4f35-b13f-c52edfbb6091","Type":"ContainerStarted","Data":"59af3c7194faf5ecc719105d47530fbaddf2a6576a6e55d1e5af4822564554d5"} Nov 22 09:13:26 crc kubenswrapper[4735]: I1122 09:13:26.753361 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rxtdz" podStartSLOduration=3.039900806 podStartE2EDuration="9.753337442s" podCreationTimestamp="2025-11-22 09:13:17 +0000 UTC" firstStartedPulling="2025-11-22 09:13:19.624015457 +0000 UTC m=+4221.228354062" lastFinishedPulling="2025-11-22 09:13:26.337452093 +0000 UTC m=+4227.941790698" observedRunningTime="2025-11-22 09:13:26.741051119 +0000 UTC m=+4228.345389724" watchObservedRunningTime="2025-11-22 09:13:26.753337442 +0000 UTC m=+4228.357676067" Nov 22 09:13:28 crc kubenswrapper[4735]: I1122 09:13:28.351992 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:28 crc kubenswrapper[4735]: I1122 09:13:28.352107 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:13:29 crc kubenswrapper[4735]: I1122 09:13:29.408274 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rxtdz" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" probeResult="failure" output=< Nov 22 09:13:29 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:13:29 crc kubenswrapper[4735]: > Nov 22 09:13:39 crc kubenswrapper[4735]: I1122 09:13:39.399570 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rxtdz" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" probeResult="failure" output=< Nov 22 09:13:39 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:13:39 crc kubenswrapper[4735]: > Nov 22 09:13:46 crc kubenswrapper[4735]: I1122 09:13:46.132321 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:13:46 crc kubenswrapper[4735]: I1122 09:13:46.133569 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:13:49 crc kubenswrapper[4735]: I1122 09:13:49.398739 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rxtdz" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" probeResult="failure" output=< Nov 22 09:13:49 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:13:49 crc kubenswrapper[4735]: > Nov 22 09:13:59 crc kubenswrapper[4735]: I1122 09:13:59.421167 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rxtdz" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" probeResult="failure" output=< Nov 22 09:13:59 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:13:59 crc kubenswrapper[4735]: > Nov 22 09:14:09 crc kubenswrapper[4735]: I1122 09:14:09.400354 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rxtdz" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" probeResult="failure" output=< Nov 22 09:14:09 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:14:09 crc kubenswrapper[4735]: > Nov 22 09:14:16 crc kubenswrapper[4735]: I1122 09:14:16.131616 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:14:16 crc kubenswrapper[4735]: I1122 09:14:16.132228 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:14:19 crc kubenswrapper[4735]: I1122 09:14:19.121153 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:14:19 crc kubenswrapper[4735]: I1122 09:14:19.207114 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:14:19 crc kubenswrapper[4735]: I1122 09:14:19.360947 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rxtdz"] Nov 22 09:14:20 crc kubenswrapper[4735]: I1122 09:14:20.314084 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rxtdz" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" containerID="cri-o://59af3c7194faf5ecc719105d47530fbaddf2a6576a6e55d1e5af4822564554d5" gracePeriod=2 Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.331276 4735 generic.go:334] "Generic (PLEG): container finished" podID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerID="59af3c7194faf5ecc719105d47530fbaddf2a6576a6e55d1e5af4822564554d5" exitCode=0 Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.331586 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxtdz" event={"ID":"07d59ecf-3c54-4f35-b13f-c52edfbb6091","Type":"ContainerDied","Data":"59af3c7194faf5ecc719105d47530fbaddf2a6576a6e55d1e5af4822564554d5"} Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.331865 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxtdz" event={"ID":"07d59ecf-3c54-4f35-b13f-c52edfbb6091","Type":"ContainerDied","Data":"b058509c31032ab22595a78445a736d2a636234b66bf15d26a1a6636932d4964"} Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.331883 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b058509c31032ab22595a78445a736d2a636234b66bf15d26a1a6636932d4964" Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.383864 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.486030 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-catalog-content\") pod \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.486104 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnv8m\" (UniqueName: \"kubernetes.io/projected/07d59ecf-3c54-4f35-b13f-c52edfbb6091-kube-api-access-rnv8m\") pod \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.486395 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-utilities\") pod \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\" (UID: \"07d59ecf-3c54-4f35-b13f-c52edfbb6091\") " Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.487098 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-utilities" (OuterVolumeSpecName: "utilities") pod "07d59ecf-3c54-4f35-b13f-c52edfbb6091" (UID: "07d59ecf-3c54-4f35-b13f-c52edfbb6091"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.493430 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07d59ecf-3c54-4f35-b13f-c52edfbb6091-kube-api-access-rnv8m" (OuterVolumeSpecName: "kube-api-access-rnv8m") pod "07d59ecf-3c54-4f35-b13f-c52edfbb6091" (UID: "07d59ecf-3c54-4f35-b13f-c52edfbb6091"). InnerVolumeSpecName "kube-api-access-rnv8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.577111 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07d59ecf-3c54-4f35-b13f-c52edfbb6091" (UID: "07d59ecf-3c54-4f35-b13f-c52edfbb6091"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.589391 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.589421 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d59ecf-3c54-4f35-b13f-c52edfbb6091-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:14:21 crc kubenswrapper[4735]: I1122 09:14:21.589431 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnv8m\" (UniqueName: \"kubernetes.io/projected/07d59ecf-3c54-4f35-b13f-c52edfbb6091-kube-api-access-rnv8m\") on node \"crc\" DevicePath \"\"" Nov 22 09:14:22 crc kubenswrapper[4735]: I1122 09:14:22.341145 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxtdz" Nov 22 09:14:22 crc kubenswrapper[4735]: I1122 09:14:22.376138 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rxtdz"] Nov 22 09:14:22 crc kubenswrapper[4735]: I1122 09:14:22.387259 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rxtdz"] Nov 22 09:14:23 crc kubenswrapper[4735]: I1122 09:14:23.280410 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" path="/var/lib/kubelet/pods/07d59ecf-3c54-4f35-b13f-c52edfbb6091/volumes" Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.131762 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.132301 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.132358 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.133360 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.133422 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" gracePeriod=600 Nov 22 09:14:46 crc kubenswrapper[4735]: E1122 09:14:46.263219 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.616107 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" exitCode=0 Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.616159 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076"} Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.616204 4735 scope.go:117] "RemoveContainer" containerID="68aa55a00176dae0321a1f4209cd4c13a08313ec37415b0b62700c8b94230560" Nov 22 09:14:46 crc kubenswrapper[4735]: I1122 09:14:46.617276 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:14:46 crc kubenswrapper[4735]: E1122 09:14:46.618053 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.278033 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vtz86"] Nov 22 09:14:47 crc kubenswrapper[4735]: E1122 09:14:47.279043 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="extract-utilities" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.279063 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="extract-utilities" Nov 22 09:14:47 crc kubenswrapper[4735]: E1122 09:14:47.279143 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="extract-content" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.279153 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="extract-content" Nov 22 09:14:47 crc kubenswrapper[4735]: E1122 09:14:47.279173 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.279181 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.279489 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="07d59ecf-3c54-4f35-b13f-c52edfbb6091" containerName="registry-server" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.281350 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.295220 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vtz86"] Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.419479 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws9mw\" (UniqueName: \"kubernetes.io/projected/e7acf3f3-53c2-40ca-8a0d-f38058f88358-kube-api-access-ws9mw\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.419594 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-catalog-content\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.419618 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-utilities\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.521712 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-catalog-content\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.521773 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-utilities\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.521948 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws9mw\" (UniqueName: \"kubernetes.io/projected/e7acf3f3-53c2-40ca-8a0d-f38058f88358-kube-api-access-ws9mw\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.522306 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-utilities\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.522306 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-catalog-content\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.540533 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws9mw\" (UniqueName: \"kubernetes.io/projected/e7acf3f3-53c2-40ca-8a0d-f38058f88358-kube-api-access-ws9mw\") pod \"community-operators-vtz86\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:47 crc kubenswrapper[4735]: I1122 09:14:47.616096 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:48 crc kubenswrapper[4735]: I1122 09:14:48.205192 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vtz86"] Nov 22 09:14:48 crc kubenswrapper[4735]: I1122 09:14:48.677770 4735 generic.go:334] "Generic (PLEG): container finished" podID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerID="27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a" exitCode=0 Nov 22 09:14:48 crc kubenswrapper[4735]: I1122 09:14:48.677818 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtz86" event={"ID":"e7acf3f3-53c2-40ca-8a0d-f38058f88358","Type":"ContainerDied","Data":"27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a"} Nov 22 09:14:48 crc kubenswrapper[4735]: I1122 09:14:48.678062 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtz86" event={"ID":"e7acf3f3-53c2-40ca-8a0d-f38058f88358","Type":"ContainerStarted","Data":"41bece16aaae9e5fdf6edba977aa06f4cbe763cd875ad61b2cbaaf870578bf72"} Nov 22 09:14:48 crc kubenswrapper[4735]: I1122 09:14:48.680490 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:14:49 crc kubenswrapper[4735]: I1122 09:14:49.692196 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtz86" event={"ID":"e7acf3f3-53c2-40ca-8a0d-f38058f88358","Type":"ContainerStarted","Data":"b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585"} Nov 22 09:14:51 crc kubenswrapper[4735]: I1122 09:14:51.718710 4735 generic.go:334] "Generic (PLEG): container finished" podID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerID="b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585" exitCode=0 Nov 22 09:14:51 crc kubenswrapper[4735]: I1122 09:14:51.718800 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtz86" event={"ID":"e7acf3f3-53c2-40ca-8a0d-f38058f88358","Type":"ContainerDied","Data":"b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585"} Nov 22 09:14:53 crc kubenswrapper[4735]: I1122 09:14:53.739573 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtz86" event={"ID":"e7acf3f3-53c2-40ca-8a0d-f38058f88358","Type":"ContainerStarted","Data":"f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae"} Nov 22 09:14:53 crc kubenswrapper[4735]: I1122 09:14:53.766864 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vtz86" podStartSLOduration=2.885995394 podStartE2EDuration="6.766845296s" podCreationTimestamp="2025-11-22 09:14:47 +0000 UTC" firstStartedPulling="2025-11-22 09:14:48.68020697 +0000 UTC m=+4310.284545575" lastFinishedPulling="2025-11-22 09:14:52.561056882 +0000 UTC m=+4314.165395477" observedRunningTime="2025-11-22 09:14:53.756939177 +0000 UTC m=+4315.361277802" watchObservedRunningTime="2025-11-22 09:14:53.766845296 +0000 UTC m=+4315.371183901" Nov 22 09:14:57 crc kubenswrapper[4735]: I1122 09:14:57.617618 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:57 crc kubenswrapper[4735]: I1122 09:14:57.618226 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:57 crc kubenswrapper[4735]: I1122 09:14:57.676081 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:57 crc kubenswrapper[4735]: I1122 09:14:57.845018 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:14:57 crc kubenswrapper[4735]: I1122 09:14:57.912860 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vtz86"] Nov 22 09:14:59 crc kubenswrapper[4735]: I1122 09:14:59.806541 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vtz86" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerName="registry-server" containerID="cri-o://f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae" gracePeriod=2 Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.172902 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm"] Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.175143 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.178910 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.179498 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.198662 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm"] Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.250802 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-secret-volume\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.250887 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-config-volume\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.250973 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqt6v\" (UniqueName: \"kubernetes.io/projected/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-kube-api-access-lqt6v\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.263934 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:15:00 crc kubenswrapper[4735]: E1122 09:15:00.264286 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.353675 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-secret-volume\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.353738 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-config-volume\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.353839 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqt6v\" (UniqueName: \"kubernetes.io/projected/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-kube-api-access-lqt6v\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.355233 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-config-volume\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.362248 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-secret-volume\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.371974 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqt6v\" (UniqueName: \"kubernetes.io/projected/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-kube-api-access-lqt6v\") pod \"collect-profiles-29396715-bvwhm\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.488755 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.509865 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.557442 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws9mw\" (UniqueName: \"kubernetes.io/projected/e7acf3f3-53c2-40ca-8a0d-f38058f88358-kube-api-access-ws9mw\") pod \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.557847 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-catalog-content\") pod \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.558010 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-utilities\") pod \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\" (UID: \"e7acf3f3-53c2-40ca-8a0d-f38058f88358\") " Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.559506 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-utilities" (OuterVolumeSpecName: "utilities") pod "e7acf3f3-53c2-40ca-8a0d-f38058f88358" (UID: "e7acf3f3-53c2-40ca-8a0d-f38058f88358"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.562639 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7acf3f3-53c2-40ca-8a0d-f38058f88358-kube-api-access-ws9mw" (OuterVolumeSpecName: "kube-api-access-ws9mw") pod "e7acf3f3-53c2-40ca-8a0d-f38058f88358" (UID: "e7acf3f3-53c2-40ca-8a0d-f38058f88358"). InnerVolumeSpecName "kube-api-access-ws9mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.611596 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7acf3f3-53c2-40ca-8a0d-f38058f88358" (UID: "e7acf3f3-53c2-40ca-8a0d-f38058f88358"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.660571 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.660605 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws9mw\" (UniqueName: \"kubernetes.io/projected/e7acf3f3-53c2-40ca-8a0d-f38058f88358-kube-api-access-ws9mw\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.660616 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7acf3f3-53c2-40ca-8a0d-f38058f88358-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.821642 4735 generic.go:334] "Generic (PLEG): container finished" podID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerID="f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae" exitCode=0 Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.821690 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtz86" event={"ID":"e7acf3f3-53c2-40ca-8a0d-f38058f88358","Type":"ContainerDied","Data":"f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae"} Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.821716 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtz86" event={"ID":"e7acf3f3-53c2-40ca-8a0d-f38058f88358","Type":"ContainerDied","Data":"41bece16aaae9e5fdf6edba977aa06f4cbe763cd875ad61b2cbaaf870578bf72"} Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.821733 4735 scope.go:117] "RemoveContainer" containerID="f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.821732 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtz86" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.851450 4735 scope.go:117] "RemoveContainer" containerID="b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.880878 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vtz86"] Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.889854 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vtz86"] Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.895783 4735 scope.go:117] "RemoveContainer" containerID="27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.918390 4735 scope.go:117] "RemoveContainer" containerID="f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae" Nov 22 09:15:00 crc kubenswrapper[4735]: E1122 09:15:00.918815 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae\": container with ID starting with f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae not found: ID does not exist" containerID="f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.918851 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae"} err="failed to get container status \"f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae\": rpc error: code = NotFound desc = could not find container \"f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae\": container with ID starting with f44c9d13966df34cd6b9f9fa95fd996ac45446d9f4e26ace8be43eb55d6facae not found: ID does not exist" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.918875 4735 scope.go:117] "RemoveContainer" containerID="b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585" Nov 22 09:15:00 crc kubenswrapper[4735]: E1122 09:15:00.919475 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585\": container with ID starting with b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585 not found: ID does not exist" containerID="b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.919534 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585"} err="failed to get container status \"b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585\": rpc error: code = NotFound desc = could not find container \"b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585\": container with ID starting with b39d75a9c7d6e70ca1cdedcb3edb3114bd3240ef736f06d31fe96eddd4b63585 not found: ID does not exist" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.919563 4735 scope.go:117] "RemoveContainer" containerID="27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a" Nov 22 09:15:00 crc kubenswrapper[4735]: E1122 09:15:00.919812 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a\": container with ID starting with 27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a not found: ID does not exist" containerID="27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a" Nov 22 09:15:00 crc kubenswrapper[4735]: I1122 09:15:00.919836 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a"} err="failed to get container status \"27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a\": rpc error: code = NotFound desc = could not find container \"27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a\": container with ID starting with 27abb23deacae89a6d7befcd984a9f84db8f61dbdcaa25afa6e9307bf3f7df4a not found: ID does not exist" Nov 22 09:15:01 crc kubenswrapper[4735]: I1122 09:15:01.036229 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm"] Nov 22 09:15:01 crc kubenswrapper[4735]: I1122 09:15:01.281997 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" path="/var/lib/kubelet/pods/e7acf3f3-53c2-40ca-8a0d-f38058f88358/volumes" Nov 22 09:15:01 crc kubenswrapper[4735]: I1122 09:15:01.833312 4735 generic.go:334] "Generic (PLEG): container finished" podID="b23b5e9d-eaad-4031-a587-45d3b0b6acb3" containerID="8d5aa322f64cb0be2209b7ee557e6186f3158bec66e9b217887f9585396a499a" exitCode=0 Nov 22 09:15:01 crc kubenswrapper[4735]: I1122 09:15:01.833393 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" event={"ID":"b23b5e9d-eaad-4031-a587-45d3b0b6acb3","Type":"ContainerDied","Data":"8d5aa322f64cb0be2209b7ee557e6186f3158bec66e9b217887f9585396a499a"} Nov 22 09:15:01 crc kubenswrapper[4735]: I1122 09:15:01.833423 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" event={"ID":"b23b5e9d-eaad-4031-a587-45d3b0b6acb3","Type":"ContainerStarted","Data":"00210d7429b95648e79f2e2b8bfb255518c00e4e7b84380b691c43878d65c682"} Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.338176 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.456334 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-config-volume\") pod \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.456428 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqt6v\" (UniqueName: \"kubernetes.io/projected/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-kube-api-access-lqt6v\") pod \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.456856 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-secret-volume\") pod \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\" (UID: \"b23b5e9d-eaad-4031-a587-45d3b0b6acb3\") " Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.457278 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-config-volume" (OuterVolumeSpecName: "config-volume") pod "b23b5e9d-eaad-4031-a587-45d3b0b6acb3" (UID: "b23b5e9d-eaad-4031-a587-45d3b0b6acb3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.458026 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.462798 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-kube-api-access-lqt6v" (OuterVolumeSpecName: "kube-api-access-lqt6v") pod "b23b5e9d-eaad-4031-a587-45d3b0b6acb3" (UID: "b23b5e9d-eaad-4031-a587-45d3b0b6acb3"). InnerVolumeSpecName "kube-api-access-lqt6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.463103 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b23b5e9d-eaad-4031-a587-45d3b0b6acb3" (UID: "b23b5e9d-eaad-4031-a587-45d3b0b6acb3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.560115 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqt6v\" (UniqueName: \"kubernetes.io/projected/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-kube-api-access-lqt6v\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.560149 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b23b5e9d-eaad-4031-a587-45d3b0b6acb3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.869638 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" event={"ID":"b23b5e9d-eaad-4031-a587-45d3b0b6acb3","Type":"ContainerDied","Data":"00210d7429b95648e79f2e2b8bfb255518c00e4e7b84380b691c43878d65c682"} Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.869985 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00210d7429b95648e79f2e2b8bfb255518c00e4e7b84380b691c43878d65c682" Nov 22 09:15:03 crc kubenswrapper[4735]: I1122 09:15:03.869721 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-bvwhm" Nov 22 09:15:04 crc kubenswrapper[4735]: I1122 09:15:04.420197 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx"] Nov 22 09:15:04 crc kubenswrapper[4735]: I1122 09:15:04.433831 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-92thx"] Nov 22 09:15:05 crc kubenswrapper[4735]: I1122 09:15:05.278205 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a12655bc-7394-4051-b783-4aa456551746" path="/var/lib/kubelet/pods/a12655bc-7394-4051-b783-4aa456551746/volumes" Nov 22 09:15:10 crc kubenswrapper[4735]: I1122 09:15:10.093209 4735 scope.go:117] "RemoveContainer" containerID="d3c8b6caf421b8b0aa649050f8c34b94e3a86cd62c43f490d99080f106efec4c" Nov 22 09:15:11 crc kubenswrapper[4735]: I1122 09:15:11.263794 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:15:11 crc kubenswrapper[4735]: E1122 09:15:11.264902 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:15:19 crc kubenswrapper[4735]: E1122 09:15:19.810073 4735 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.234:59860->38.102.83.234:33833: write tcp 38.102.83.234:59860->38.102.83.234:33833: write: broken pipe Nov 22 09:15:22 crc kubenswrapper[4735]: I1122 09:15:22.265037 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:15:22 crc kubenswrapper[4735]: E1122 09:15:22.266020 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:15:36 crc kubenswrapper[4735]: I1122 09:15:36.264405 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:15:36 crc kubenswrapper[4735]: E1122 09:15:36.265261 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:15:48 crc kubenswrapper[4735]: I1122 09:15:48.263821 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:15:48 crc kubenswrapper[4735]: E1122 09:15:48.264939 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:16:00 crc kubenswrapper[4735]: I1122 09:16:00.263309 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:16:00 crc kubenswrapper[4735]: E1122 09:16:00.264023 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:16:11 crc kubenswrapper[4735]: I1122 09:16:11.263427 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:16:11 crc kubenswrapper[4735]: E1122 09:16:11.264803 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:16:25 crc kubenswrapper[4735]: I1122 09:16:25.264528 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:16:25 crc kubenswrapper[4735]: E1122 09:16:25.266708 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:16:39 crc kubenswrapper[4735]: I1122 09:16:39.271292 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:16:39 crc kubenswrapper[4735]: E1122 09:16:39.272125 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:16:52 crc kubenswrapper[4735]: I1122 09:16:52.264781 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:16:52 crc kubenswrapper[4735]: E1122 09:16:52.267219 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:17:07 crc kubenswrapper[4735]: I1122 09:17:07.263497 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:17:07 crc kubenswrapper[4735]: E1122 09:17:07.264417 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:17:19 crc kubenswrapper[4735]: I1122 09:17:19.271713 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:17:19 crc kubenswrapper[4735]: E1122 09:17:19.272610 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:17:30 crc kubenswrapper[4735]: I1122 09:17:30.264537 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:17:30 crc kubenswrapper[4735]: E1122 09:17:30.265303 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:17:44 crc kubenswrapper[4735]: I1122 09:17:44.263777 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:17:44 crc kubenswrapper[4735]: E1122 09:17:44.264563 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:17:59 crc kubenswrapper[4735]: I1122 09:17:59.274850 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:17:59 crc kubenswrapper[4735]: E1122 09:17:59.277067 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:18:10 crc kubenswrapper[4735]: I1122 09:18:10.264378 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:18:10 crc kubenswrapper[4735]: E1122 09:18:10.265391 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:18:23 crc kubenswrapper[4735]: I1122 09:18:23.271211 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:18:23 crc kubenswrapper[4735]: E1122 09:18:23.272014 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:18:38 crc kubenswrapper[4735]: I1122 09:18:38.264277 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:18:38 crc kubenswrapper[4735]: E1122 09:18:38.264966 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:18:49 crc kubenswrapper[4735]: I1122 09:18:49.275558 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:18:49 crc kubenswrapper[4735]: E1122 09:18:49.276243 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:19:04 crc kubenswrapper[4735]: I1122 09:19:04.264115 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:19:04 crc kubenswrapper[4735]: E1122 09:19:04.264848 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:19:17 crc kubenswrapper[4735]: I1122 09:19:17.263186 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:19:17 crc kubenswrapper[4735]: E1122 09:19:17.264258 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.336974 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sn888"] Nov 22 09:19:29 crc kubenswrapper[4735]: E1122 09:19:29.337886 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerName="extract-content" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.337899 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerName="extract-content" Nov 22 09:19:29 crc kubenswrapper[4735]: E1122 09:19:29.337940 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b23b5e9d-eaad-4031-a587-45d3b0b6acb3" containerName="collect-profiles" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.337946 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="b23b5e9d-eaad-4031-a587-45d3b0b6acb3" containerName="collect-profiles" Nov 22 09:19:29 crc kubenswrapper[4735]: E1122 09:19:29.337972 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerName="registry-server" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.337978 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerName="registry-server" Nov 22 09:19:29 crc kubenswrapper[4735]: E1122 09:19:29.337989 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerName="extract-utilities" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.337995 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerName="extract-utilities" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.338231 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7acf3f3-53c2-40ca-8a0d-f38058f88358" containerName="registry-server" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.338247 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="b23b5e9d-eaad-4031-a587-45d3b0b6acb3" containerName="collect-profiles" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.339979 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.355231 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sn888"] Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.444059 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-utilities\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.444498 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdvmd\" (UniqueName: \"kubernetes.io/projected/ac7a7e00-b385-4157-b88d-7c33b644b2c2-kube-api-access-sdvmd\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.444613 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-catalog-content\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.546934 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-utilities\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.547072 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdvmd\" (UniqueName: \"kubernetes.io/projected/ac7a7e00-b385-4157-b88d-7c33b644b2c2-kube-api-access-sdvmd\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.547179 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-catalog-content\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.547515 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-utilities\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.547653 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-catalog-content\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.571481 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdvmd\" (UniqueName: \"kubernetes.io/projected/ac7a7e00-b385-4157-b88d-7c33b644b2c2-kube-api-access-sdvmd\") pod \"redhat-marketplace-sn888\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:29 crc kubenswrapper[4735]: I1122 09:19:29.667309 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:30 crc kubenswrapper[4735]: I1122 09:19:30.135184 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sn888"] Nov 22 09:19:30 crc kubenswrapper[4735]: I1122 09:19:30.846000 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sn888" event={"ID":"ac7a7e00-b385-4157-b88d-7c33b644b2c2","Type":"ContainerStarted","Data":"fb73c9a5540a6b9e58273924591f1d80346dc371ba354143afbca0338386e43a"} Nov 22 09:19:31 crc kubenswrapper[4735]: I1122 09:19:31.263472 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:19:31 crc kubenswrapper[4735]: E1122 09:19:31.264041 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:19:31 crc kubenswrapper[4735]: I1122 09:19:31.859995 4735 generic.go:334] "Generic (PLEG): container finished" podID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerID="d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6" exitCode=0 Nov 22 09:19:31 crc kubenswrapper[4735]: I1122 09:19:31.860136 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sn888" event={"ID":"ac7a7e00-b385-4157-b88d-7c33b644b2c2","Type":"ContainerDied","Data":"d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6"} Nov 22 09:19:33 crc kubenswrapper[4735]: I1122 09:19:33.891133 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sn888" event={"ID":"ac7a7e00-b385-4157-b88d-7c33b644b2c2","Type":"ContainerStarted","Data":"2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69"} Nov 22 09:19:34 crc kubenswrapper[4735]: I1122 09:19:34.907573 4735 generic.go:334] "Generic (PLEG): container finished" podID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerID="2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69" exitCode=0 Nov 22 09:19:34 crc kubenswrapper[4735]: I1122 09:19:34.907675 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sn888" event={"ID":"ac7a7e00-b385-4157-b88d-7c33b644b2c2","Type":"ContainerDied","Data":"2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69"} Nov 22 09:19:35 crc kubenswrapper[4735]: I1122 09:19:35.922295 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sn888" event={"ID":"ac7a7e00-b385-4157-b88d-7c33b644b2c2","Type":"ContainerStarted","Data":"cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6"} Nov 22 09:19:35 crc kubenswrapper[4735]: I1122 09:19:35.951810 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sn888" podStartSLOduration=3.457936524 podStartE2EDuration="6.951784655s" podCreationTimestamp="2025-11-22 09:19:29 +0000 UTC" firstStartedPulling="2025-11-22 09:19:31.862376245 +0000 UTC m=+4593.466714850" lastFinishedPulling="2025-11-22 09:19:35.356224376 +0000 UTC m=+4596.960562981" observedRunningTime="2025-11-22 09:19:35.938262428 +0000 UTC m=+4597.542601033" watchObservedRunningTime="2025-11-22 09:19:35.951784655 +0000 UTC m=+4597.556123270" Nov 22 09:19:39 crc kubenswrapper[4735]: I1122 09:19:39.667952 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:39 crc kubenswrapper[4735]: I1122 09:19:39.668487 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:39 crc kubenswrapper[4735]: I1122 09:19:39.734107 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:45 crc kubenswrapper[4735]: I1122 09:19:45.264316 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:19:45 crc kubenswrapper[4735]: E1122 09:19:45.265157 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:19:49 crc kubenswrapper[4735]: I1122 09:19:49.726098 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:49 crc kubenswrapper[4735]: I1122 09:19:49.789257 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sn888"] Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.072864 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sn888" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerName="registry-server" containerID="cri-o://cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6" gracePeriod=2 Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.635091 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.753602 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdvmd\" (UniqueName: \"kubernetes.io/projected/ac7a7e00-b385-4157-b88d-7c33b644b2c2-kube-api-access-sdvmd\") pod \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.753679 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-catalog-content\") pod \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.753907 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-utilities\") pod \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\" (UID: \"ac7a7e00-b385-4157-b88d-7c33b644b2c2\") " Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.754513 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-utilities" (OuterVolumeSpecName: "utilities") pod "ac7a7e00-b385-4157-b88d-7c33b644b2c2" (UID: "ac7a7e00-b385-4157-b88d-7c33b644b2c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.754952 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.760211 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac7a7e00-b385-4157-b88d-7c33b644b2c2-kube-api-access-sdvmd" (OuterVolumeSpecName: "kube-api-access-sdvmd") pod "ac7a7e00-b385-4157-b88d-7c33b644b2c2" (UID: "ac7a7e00-b385-4157-b88d-7c33b644b2c2"). InnerVolumeSpecName "kube-api-access-sdvmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.819611 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac7a7e00-b385-4157-b88d-7c33b644b2c2" (UID: "ac7a7e00-b385-4157-b88d-7c33b644b2c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.857111 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdvmd\" (UniqueName: \"kubernetes.io/projected/ac7a7e00-b385-4157-b88d-7c33b644b2c2-kube-api-access-sdvmd\") on node \"crc\" DevicePath \"\"" Nov 22 09:19:50 crc kubenswrapper[4735]: I1122 09:19:50.857367 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7a7e00-b385-4157-b88d-7c33b644b2c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.087673 4735 generic.go:334] "Generic (PLEG): container finished" podID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerID="cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6" exitCode=0 Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.087713 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sn888" event={"ID":"ac7a7e00-b385-4157-b88d-7c33b644b2c2","Type":"ContainerDied","Data":"cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6"} Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.087738 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sn888" event={"ID":"ac7a7e00-b385-4157-b88d-7c33b644b2c2","Type":"ContainerDied","Data":"fb73c9a5540a6b9e58273924591f1d80346dc371ba354143afbca0338386e43a"} Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.087757 4735 scope.go:117] "RemoveContainer" containerID="cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.087882 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sn888" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.125826 4735 scope.go:117] "RemoveContainer" containerID="2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.129711 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sn888"] Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.149286 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sn888"] Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.159742 4735 scope.go:117] "RemoveContainer" containerID="d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.219322 4735 scope.go:117] "RemoveContainer" containerID="cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6" Nov 22 09:19:51 crc kubenswrapper[4735]: E1122 09:19:51.219748 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6\": container with ID starting with cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6 not found: ID does not exist" containerID="cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.219784 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6"} err="failed to get container status \"cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6\": rpc error: code = NotFound desc = could not find container \"cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6\": container with ID starting with cc7321cc5390162953630f7a20add8337db9de99e5d3ede66dbb017ee26bdeb6 not found: ID does not exist" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.219831 4735 scope.go:117] "RemoveContainer" containerID="2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69" Nov 22 09:19:51 crc kubenswrapper[4735]: E1122 09:19:51.220099 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69\": container with ID starting with 2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69 not found: ID does not exist" containerID="2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.220124 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69"} err="failed to get container status \"2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69\": rpc error: code = NotFound desc = could not find container \"2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69\": container with ID starting with 2d23e9ffcbe49d42e8523c9b0de7c2eae676c527260bcec03c61a30f28981f69 not found: ID does not exist" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.220143 4735 scope.go:117] "RemoveContainer" containerID="d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6" Nov 22 09:19:51 crc kubenswrapper[4735]: E1122 09:19:51.220390 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6\": container with ID starting with d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6 not found: ID does not exist" containerID="d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.220411 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6"} err="failed to get container status \"d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6\": rpc error: code = NotFound desc = could not find container \"d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6\": container with ID starting with d57aa1aa8cd9fa0f6df369e652d42599e5a715eb2dfc5d27aa775189783e26a6 not found: ID does not exist" Nov 22 09:19:51 crc kubenswrapper[4735]: I1122 09:19:51.278608 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" path="/var/lib/kubelet/pods/ac7a7e00-b385-4157-b88d-7c33b644b2c2/volumes" Nov 22 09:20:00 crc kubenswrapper[4735]: I1122 09:20:00.263257 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:20:01 crc kubenswrapper[4735]: I1122 09:20:01.187782 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"18f3c56d6b82827ef0d9798d93b1e5c6a5288d3b09223b912c74cba8a8ae1f73"} Nov 22 09:20:10 crc kubenswrapper[4735]: I1122 09:20:10.293857 4735 scope.go:117] "RemoveContainer" containerID="59af3c7194faf5ecc719105d47530fbaddf2a6576a6e55d1e5af4822564554d5" Nov 22 09:20:10 crc kubenswrapper[4735]: I1122 09:20:10.317566 4735 scope.go:117] "RemoveContainer" containerID="27f7de275e90d12fdbdeacb62d87b0d37ed4fd839ad1c1dd5953dd7887684338" Nov 22 09:20:10 crc kubenswrapper[4735]: I1122 09:20:10.339337 4735 scope.go:117] "RemoveContainer" containerID="e1eef004b1c2d6c94c7362c66bc5afe2e951d13d214d58b3ed361c0a888a2e0f" Nov 22 09:22:16 crc kubenswrapper[4735]: I1122 09:22:16.131541 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:22:16 crc kubenswrapper[4735]: I1122 09:22:16.132130 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.406133 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ddbr6"] Nov 22 09:22:21 crc kubenswrapper[4735]: E1122 09:22:21.406948 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerName="extract-content" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.406964 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerName="extract-content" Nov 22 09:22:21 crc kubenswrapper[4735]: E1122 09:22:21.406995 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerName="registry-server" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.407000 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerName="registry-server" Nov 22 09:22:21 crc kubenswrapper[4735]: E1122 09:22:21.407019 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerName="extract-utilities" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.407026 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerName="extract-utilities" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.407241 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac7a7e00-b385-4157-b88d-7c33b644b2c2" containerName="registry-server" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.409146 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.434668 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ddbr6"] Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.506212 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-catalog-content\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.506508 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq9kn\" (UniqueName: \"kubernetes.io/projected/7a030cfc-d64c-458a-85e9-fd660f223dd9-kube-api-access-lq9kn\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.506685 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-utilities\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.609032 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq9kn\" (UniqueName: \"kubernetes.io/projected/7a030cfc-d64c-458a-85e9-fd660f223dd9-kube-api-access-lq9kn\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.609111 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-utilities\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.609231 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-catalog-content\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.609818 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-catalog-content\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.609826 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-utilities\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.637628 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq9kn\" (UniqueName: \"kubernetes.io/projected/7a030cfc-d64c-458a-85e9-fd660f223dd9-kube-api-access-lq9kn\") pod \"certified-operators-ddbr6\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:21 crc kubenswrapper[4735]: I1122 09:22:21.732052 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:22 crc kubenswrapper[4735]: I1122 09:22:22.257679 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ddbr6"] Nov 22 09:22:22 crc kubenswrapper[4735]: I1122 09:22:22.703389 4735 generic.go:334] "Generic (PLEG): container finished" podID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerID="1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6" exitCode=0 Nov 22 09:22:22 crc kubenswrapper[4735]: I1122 09:22:22.703704 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddbr6" event={"ID":"7a030cfc-d64c-458a-85e9-fd660f223dd9","Type":"ContainerDied","Data":"1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6"} Nov 22 09:22:22 crc kubenswrapper[4735]: I1122 09:22:22.703731 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddbr6" event={"ID":"7a030cfc-d64c-458a-85e9-fd660f223dd9","Type":"ContainerStarted","Data":"40fefa4d399b17699d58bff3037018abe471896d24835d5bd8904b35bbcffc0d"} Nov 22 09:22:22 crc kubenswrapper[4735]: I1122 09:22:22.705690 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:22:23 crc kubenswrapper[4735]: I1122 09:22:23.716109 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddbr6" event={"ID":"7a030cfc-d64c-458a-85e9-fd660f223dd9","Type":"ContainerStarted","Data":"cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844"} Nov 22 09:22:25 crc kubenswrapper[4735]: I1122 09:22:25.738693 4735 generic.go:334] "Generic (PLEG): container finished" podID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerID="cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844" exitCode=0 Nov 22 09:22:25 crc kubenswrapper[4735]: I1122 09:22:25.738803 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddbr6" event={"ID":"7a030cfc-d64c-458a-85e9-fd660f223dd9","Type":"ContainerDied","Data":"cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844"} Nov 22 09:22:26 crc kubenswrapper[4735]: I1122 09:22:26.751357 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddbr6" event={"ID":"7a030cfc-d64c-458a-85e9-fd660f223dd9","Type":"ContainerStarted","Data":"09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239"} Nov 22 09:22:26 crc kubenswrapper[4735]: I1122 09:22:26.778706 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ddbr6" podStartSLOduration=2.319040909 podStartE2EDuration="5.778682813s" podCreationTimestamp="2025-11-22 09:22:21 +0000 UTC" firstStartedPulling="2025-11-22 09:22:22.705438443 +0000 UTC m=+4764.309777048" lastFinishedPulling="2025-11-22 09:22:26.165080357 +0000 UTC m=+4767.769418952" observedRunningTime="2025-11-22 09:22:26.767805967 +0000 UTC m=+4768.372144572" watchObservedRunningTime="2025-11-22 09:22:26.778682813 +0000 UTC m=+4768.383021418" Nov 22 09:22:31 crc kubenswrapper[4735]: I1122 09:22:31.732399 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:31 crc kubenswrapper[4735]: I1122 09:22:31.732721 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:31 crc kubenswrapper[4735]: I1122 09:22:31.786875 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:31 crc kubenswrapper[4735]: I1122 09:22:31.857176 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:32 crc kubenswrapper[4735]: I1122 09:22:32.594651 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ddbr6"] Nov 22 09:22:33 crc kubenswrapper[4735]: I1122 09:22:33.825029 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ddbr6" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerName="registry-server" containerID="cri-o://09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239" gracePeriod=2 Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.435643 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.547488 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-catalog-content\") pod \"7a030cfc-d64c-458a-85e9-fd660f223dd9\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.547525 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-utilities\") pod \"7a030cfc-d64c-458a-85e9-fd660f223dd9\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.547573 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq9kn\" (UniqueName: \"kubernetes.io/projected/7a030cfc-d64c-458a-85e9-fd660f223dd9-kube-api-access-lq9kn\") pod \"7a030cfc-d64c-458a-85e9-fd660f223dd9\" (UID: \"7a030cfc-d64c-458a-85e9-fd660f223dd9\") " Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.548543 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-utilities" (OuterVolumeSpecName: "utilities") pod "7a030cfc-d64c-458a-85e9-fd660f223dd9" (UID: "7a030cfc-d64c-458a-85e9-fd660f223dd9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.556173 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a030cfc-d64c-458a-85e9-fd660f223dd9-kube-api-access-lq9kn" (OuterVolumeSpecName: "kube-api-access-lq9kn") pod "7a030cfc-d64c-458a-85e9-fd660f223dd9" (UID: "7a030cfc-d64c-458a-85e9-fd660f223dd9"). InnerVolumeSpecName "kube-api-access-lq9kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.594691 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a030cfc-d64c-458a-85e9-fd660f223dd9" (UID: "7a030cfc-d64c-458a-85e9-fd660f223dd9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.650490 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.650538 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a030cfc-d64c-458a-85e9-fd660f223dd9-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.650552 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq9kn\" (UniqueName: \"kubernetes.io/projected/7a030cfc-d64c-458a-85e9-fd660f223dd9-kube-api-access-lq9kn\") on node \"crc\" DevicePath \"\"" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.837341 4735 generic.go:334] "Generic (PLEG): container finished" podID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerID="09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239" exitCode=0 Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.837391 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddbr6" event={"ID":"7a030cfc-d64c-458a-85e9-fd660f223dd9","Type":"ContainerDied","Data":"09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239"} Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.837422 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddbr6" event={"ID":"7a030cfc-d64c-458a-85e9-fd660f223dd9","Type":"ContainerDied","Data":"40fefa4d399b17699d58bff3037018abe471896d24835d5bd8904b35bbcffc0d"} Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.837445 4735 scope.go:117] "RemoveContainer" containerID="09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.839560 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddbr6" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.861735 4735 scope.go:117] "RemoveContainer" containerID="cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.877911 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ddbr6"] Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.892008 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ddbr6"] Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.904542 4735 scope.go:117] "RemoveContainer" containerID="1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.941587 4735 scope.go:117] "RemoveContainer" containerID="09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239" Nov 22 09:22:34 crc kubenswrapper[4735]: E1122 09:22:34.942008 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239\": container with ID starting with 09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239 not found: ID does not exist" containerID="09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.942053 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239"} err="failed to get container status \"09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239\": rpc error: code = NotFound desc = could not find container \"09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239\": container with ID starting with 09cc1891e7e34778fd3f1a82df256649d14fa0bb6d8007ce6fe73d6437c33239 not found: ID does not exist" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.942083 4735 scope.go:117] "RemoveContainer" containerID="cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844" Nov 22 09:22:34 crc kubenswrapper[4735]: E1122 09:22:34.942345 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844\": container with ID starting with cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844 not found: ID does not exist" containerID="cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.942375 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844"} err="failed to get container status \"cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844\": rpc error: code = NotFound desc = could not find container \"cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844\": container with ID starting with cf376b166e82f0b9de79e68609b4d54d6251b1adaf3ac044e4e7a69551025844 not found: ID does not exist" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.942395 4735 scope.go:117] "RemoveContainer" containerID="1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6" Nov 22 09:22:34 crc kubenswrapper[4735]: E1122 09:22:34.942600 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6\": container with ID starting with 1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6 not found: ID does not exist" containerID="1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6" Nov 22 09:22:34 crc kubenswrapper[4735]: I1122 09:22:34.942622 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6"} err="failed to get container status \"1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6\": rpc error: code = NotFound desc = could not find container \"1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6\": container with ID starting with 1822c0b1530fe457bfc63668a51ae7a4f51836c93cd225d9cd316c1d4d57e7b6 not found: ID does not exist" Nov 22 09:22:35 crc kubenswrapper[4735]: I1122 09:22:35.280351 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" path="/var/lib/kubelet/pods/7a030cfc-d64c-458a-85e9-fd660f223dd9/volumes" Nov 22 09:22:46 crc kubenswrapper[4735]: I1122 09:22:46.131105 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:22:46 crc kubenswrapper[4735]: I1122 09:22:46.131674 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.159185 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 22 09:23:14 crc kubenswrapper[4735]: E1122 09:23:14.160408 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerName="extract-utilities" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.160429 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerName="extract-utilities" Nov 22 09:23:14 crc kubenswrapper[4735]: E1122 09:23:14.160453 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerName="extract-content" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.160483 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerName="extract-content" Nov 22 09:23:14 crc kubenswrapper[4735]: E1122 09:23:14.160498 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerName="registry-server" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.160506 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerName="registry-server" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.160856 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a030cfc-d64c-458a-85e9-fd660f223dd9" containerName="registry-server" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.162002 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.163876 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.164542 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.164699 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.165295 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-2lmrg" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.173564 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268269 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268358 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v72xk\" (UniqueName: \"kubernetes.io/projected/9d89aa19-7bdb-4db2-ad64-438da4087c36-kube-api-access-v72xk\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268416 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-config-data\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268611 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268634 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268739 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268807 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268841 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.268900 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.371605 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.371663 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v72xk\" (UniqueName: \"kubernetes.io/projected/9d89aa19-7bdb-4db2-ad64-438da4087c36-kube-api-access-v72xk\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.371724 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-config-data\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.371871 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.371893 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.371935 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.371959 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.371981 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.372015 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.372866 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.373497 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.373914 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.374201 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-config-data\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.377560 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.380989 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.381180 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.381522 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.406732 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v72xk\" (UniqueName: \"kubernetes.io/projected/9d89aa19-7bdb-4db2-ad64-438da4087c36-kube-api-access-v72xk\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.444438 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.490102 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 22 09:23:14 crc kubenswrapper[4735]: I1122 09:23:14.981604 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 22 09:23:15 crc kubenswrapper[4735]: I1122 09:23:15.423888 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9d89aa19-7bdb-4db2-ad64-438da4087c36","Type":"ContainerStarted","Data":"1a43c20d6071e440ed3eb2a2dc72346e6ca3feb2cfbc7219f2d9ceccebbf61be"} Nov 22 09:23:16 crc kubenswrapper[4735]: I1122 09:23:16.131423 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:23:16 crc kubenswrapper[4735]: I1122 09:23:16.131838 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:23:16 crc kubenswrapper[4735]: I1122 09:23:16.131929 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:23:16 crc kubenswrapper[4735]: I1122 09:23:16.132894 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"18f3c56d6b82827ef0d9798d93b1e5c6a5288d3b09223b912c74cba8a8ae1f73"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:23:16 crc kubenswrapper[4735]: I1122 09:23:16.132975 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://18f3c56d6b82827ef0d9798d93b1e5c6a5288d3b09223b912c74cba8a8ae1f73" gracePeriod=600 Nov 22 09:23:16 crc kubenswrapper[4735]: I1122 09:23:16.445484 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="18f3c56d6b82827ef0d9798d93b1e5c6a5288d3b09223b912c74cba8a8ae1f73" exitCode=0 Nov 22 09:23:16 crc kubenswrapper[4735]: I1122 09:23:16.445724 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"18f3c56d6b82827ef0d9798d93b1e5c6a5288d3b09223b912c74cba8a8ae1f73"} Nov 22 09:23:16 crc kubenswrapper[4735]: I1122 09:23:16.445953 4735 scope.go:117] "RemoveContainer" containerID="d2ee2193e763b6df79fceee49934ecc6ec51fe207c353fd36bffe2abbbd92076" Nov 22 09:23:17 crc kubenswrapper[4735]: I1122 09:23:17.459091 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7"} Nov 22 09:23:48 crc kubenswrapper[4735]: E1122 09:23:48.446947 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 22 09:23:48 crc kubenswrapper[4735]: E1122 09:23:48.451411 4735 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v72xk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(9d89aa19-7bdb-4db2-ad64-438da4087c36): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 09:23:48 crc kubenswrapper[4735]: E1122 09:23:48.452658 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="9d89aa19-7bdb-4db2-ad64-438da4087c36" Nov 22 09:23:48 crc kubenswrapper[4735]: E1122 09:23:48.810957 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="9d89aa19-7bdb-4db2-ad64-438da4087c36" Nov 22 09:24:03 crc kubenswrapper[4735]: I1122 09:24:03.709336 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 22 09:24:06 crc kubenswrapper[4735]: I1122 09:24:06.031246 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9d89aa19-7bdb-4db2-ad64-438da4087c36","Type":"ContainerStarted","Data":"a4f2b24b4ad681dbaf10f89723ebe5e7d6b16ef9244464d9351710b372ccee1a"} Nov 22 09:24:06 crc kubenswrapper[4735]: I1122 09:24:06.055675 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.3465175 podStartE2EDuration="53.055652087s" podCreationTimestamp="2025-11-22 09:23:13 +0000 UTC" firstStartedPulling="2025-11-22 09:23:14.996086097 +0000 UTC m=+4816.600424702" lastFinishedPulling="2025-11-22 09:24:03.705220684 +0000 UTC m=+4865.309559289" observedRunningTime="2025-11-22 09:24:06.051016479 +0000 UTC m=+4867.655355094" watchObservedRunningTime="2025-11-22 09:24:06.055652087 +0000 UTC m=+4867.659990702" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.570742 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-572lq"] Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.576777 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.627857 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-catalog-content\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.628512 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-utilities\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.628685 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7pnf\" (UniqueName: \"kubernetes.io/projected/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-kube-api-access-j7pnf\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.654545 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-572lq"] Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.732926 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-catalog-content\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.733065 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-utilities\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.733099 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7pnf\" (UniqueName: \"kubernetes.io/projected/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-kube-api-access-j7pnf\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.735572 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-catalog-content\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.736095 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-utilities\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.760598 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7pnf\" (UniqueName: \"kubernetes.io/projected/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-kube-api-access-j7pnf\") pod \"redhat-operators-572lq\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:57 crc kubenswrapper[4735]: I1122 09:24:57.900607 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:24:59 crc kubenswrapper[4735]: I1122 09:24:59.088134 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-572lq"] Nov 22 09:24:59 crc kubenswrapper[4735]: I1122 09:24:59.616960 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-572lq" event={"ID":"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba","Type":"ContainerDied","Data":"5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9"} Nov 22 09:24:59 crc kubenswrapper[4735]: I1122 09:24:59.616857 4735 generic.go:334] "Generic (PLEG): container finished" podID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerID="5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9" exitCode=0 Nov 22 09:24:59 crc kubenswrapper[4735]: I1122 09:24:59.617343 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-572lq" event={"ID":"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba","Type":"ContainerStarted","Data":"8c1beefcd652f8f577c6af9c9da9912518329ba9c5993bc74b064cea05898557"} Nov 22 09:25:00 crc kubenswrapper[4735]: I1122 09:25:00.669710 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-572lq" event={"ID":"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba","Type":"ContainerStarted","Data":"32261606ce08528586b4beeabbbfe39eec3624f380b1addc7d9a5ba07611dd5b"} Nov 22 09:25:00 crc kubenswrapper[4735]: I1122 09:25:00.926181 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gvx69"] Nov 22 09:25:00 crc kubenswrapper[4735]: I1122 09:25:00.929097 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:00 crc kubenswrapper[4735]: I1122 09:25:00.943498 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gvx69"] Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.014721 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmvhs\" (UniqueName: \"kubernetes.io/projected/0311e0ca-8871-4ca4-8123-138ec35497c5-kube-api-access-cmvhs\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.014977 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-utilities\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.015179 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-catalog-content\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.117353 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-utilities\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.117445 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-catalog-content\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.117689 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmvhs\" (UniqueName: \"kubernetes.io/projected/0311e0ca-8871-4ca4-8123-138ec35497c5-kube-api-access-cmvhs\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.134012 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-utilities\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.134099 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-catalog-content\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.356803 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmvhs\" (UniqueName: \"kubernetes.io/projected/0311e0ca-8871-4ca4-8123-138ec35497c5-kube-api-access-cmvhs\") pod \"community-operators-gvx69\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:01 crc kubenswrapper[4735]: I1122 09:25:01.548842 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:02 crc kubenswrapper[4735]: I1122 09:25:02.121795 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gvx69"] Nov 22 09:25:02 crc kubenswrapper[4735]: W1122 09:25:02.171219 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0311e0ca_8871_4ca4_8123_138ec35497c5.slice/crio-c3fdc699d31fea8f8f96c647ac584a8ce64b0b1ed4e6f4f83f015637d4dbcb3f WatchSource:0}: Error finding container c3fdc699d31fea8f8f96c647ac584a8ce64b0b1ed4e6f4f83f015637d4dbcb3f: Status 404 returned error can't find the container with id c3fdc699d31fea8f8f96c647ac584a8ce64b0b1ed4e6f4f83f015637d4dbcb3f Nov 22 09:25:02 crc kubenswrapper[4735]: I1122 09:25:02.710360 4735 generic.go:334] "Generic (PLEG): container finished" podID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerID="cf7c4fad2b8bacd7345143f0aa131cd37b69377e7053e6fe42bd86b92020d087" exitCode=0 Nov 22 09:25:02 crc kubenswrapper[4735]: I1122 09:25:02.710633 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvx69" event={"ID":"0311e0ca-8871-4ca4-8123-138ec35497c5","Type":"ContainerDied","Data":"cf7c4fad2b8bacd7345143f0aa131cd37b69377e7053e6fe42bd86b92020d087"} Nov 22 09:25:02 crc kubenswrapper[4735]: I1122 09:25:02.710659 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvx69" event={"ID":"0311e0ca-8871-4ca4-8123-138ec35497c5","Type":"ContainerStarted","Data":"c3fdc699d31fea8f8f96c647ac584a8ce64b0b1ed4e6f4f83f015637d4dbcb3f"} Nov 22 09:25:04 crc kubenswrapper[4735]: I1122 09:25:04.741349 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvx69" event={"ID":"0311e0ca-8871-4ca4-8123-138ec35497c5","Type":"ContainerStarted","Data":"797b0cc556e53d4a14a5a0c2facb4808cec92e4b48a95485d4c45ce8040c55cc"} Nov 22 09:25:06 crc kubenswrapper[4735]: E1122 09:25:06.151802 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:07 crc kubenswrapper[4735]: I1122 09:25:07.780318 4735 generic.go:334] "Generic (PLEG): container finished" podID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerID="32261606ce08528586b4beeabbbfe39eec3624f380b1addc7d9a5ba07611dd5b" exitCode=0 Nov 22 09:25:07 crc kubenswrapper[4735]: I1122 09:25:07.780407 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-572lq" event={"ID":"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba","Type":"ContainerDied","Data":"32261606ce08528586b4beeabbbfe39eec3624f380b1addc7d9a5ba07611dd5b"} Nov 22 09:25:08 crc kubenswrapper[4735]: E1122 09:25:08.390494 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:09 crc kubenswrapper[4735]: I1122 09:25:09.808475 4735 generic.go:334] "Generic (PLEG): container finished" podID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerID="797b0cc556e53d4a14a5a0c2facb4808cec92e4b48a95485d4c45ce8040c55cc" exitCode=0 Nov 22 09:25:09 crc kubenswrapper[4735]: I1122 09:25:09.808717 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvx69" event={"ID":"0311e0ca-8871-4ca4-8123-138ec35497c5","Type":"ContainerDied","Data":"797b0cc556e53d4a14a5a0c2facb4808cec92e4b48a95485d4c45ce8040c55cc"} Nov 22 09:25:10 crc kubenswrapper[4735]: I1122 09:25:10.822916 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvx69" event={"ID":"0311e0ca-8871-4ca4-8123-138ec35497c5","Type":"ContainerStarted","Data":"1c51b907d875a6e5568da8f77da38f59223e824a21a8d04e68d66f0c3432f959"} Nov 22 09:25:10 crc kubenswrapper[4735]: I1122 09:25:10.826445 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-572lq" event={"ID":"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba","Type":"ContainerStarted","Data":"c8d02a598743ac330c998d1451ffd32de338ea9dcbe283271f6bb0ddc9ac04e4"} Nov 22 09:25:10 crc kubenswrapper[4735]: I1122 09:25:10.855640 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gvx69" podStartSLOduration=3.330494253 podStartE2EDuration="10.847778773s" podCreationTimestamp="2025-11-22 09:25:00 +0000 UTC" firstStartedPulling="2025-11-22 09:25:02.712273898 +0000 UTC m=+4924.316612503" lastFinishedPulling="2025-11-22 09:25:10.229558418 +0000 UTC m=+4931.833897023" observedRunningTime="2025-11-22 09:25:10.843782263 +0000 UTC m=+4932.448120878" watchObservedRunningTime="2025-11-22 09:25:10.847778773 +0000 UTC m=+4932.452117378" Nov 22 09:25:10 crc kubenswrapper[4735]: I1122 09:25:10.876087 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-572lq" podStartSLOduration=3.653443223 podStartE2EDuration="13.876064029s" podCreationTimestamp="2025-11-22 09:24:57 +0000 UTC" firstStartedPulling="2025-11-22 09:24:59.619479515 +0000 UTC m=+4921.223818120" lastFinishedPulling="2025-11-22 09:25:09.842100321 +0000 UTC m=+4931.446438926" observedRunningTime="2025-11-22 09:25:10.86478427 +0000 UTC m=+4932.469122895" watchObservedRunningTime="2025-11-22 09:25:10.876064029 +0000 UTC m=+4932.480402634" Nov 22 09:25:11 crc kubenswrapper[4735]: I1122 09:25:11.549863 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:11 crc kubenswrapper[4735]: I1122 09:25:11.550205 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:12 crc kubenswrapper[4735]: I1122 09:25:12.611354 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gvx69" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="registry-server" probeResult="failure" output=< Nov 22 09:25:12 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:25:12 crc kubenswrapper[4735]: > Nov 22 09:25:16 crc kubenswrapper[4735]: I1122 09:25:16.145702 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:25:16 crc kubenswrapper[4735]: I1122 09:25:16.151762 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:25:16 crc kubenswrapper[4735]: E1122 09:25:16.539920 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:17 crc kubenswrapper[4735]: I1122 09:25:17.901265 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:25:17 crc kubenswrapper[4735]: I1122 09:25:17.901779 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:25:18 crc kubenswrapper[4735]: I1122 09:25:18.994519 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-572lq" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" probeResult="failure" output=< Nov 22 09:25:18 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:25:18 crc kubenswrapper[4735]: > Nov 22 09:25:22 crc kubenswrapper[4735]: I1122 09:25:22.616838 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gvx69" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="registry-server" probeResult="failure" output=< Nov 22 09:25:22 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:25:22 crc kubenswrapper[4735]: > Nov 22 09:25:23 crc kubenswrapper[4735]: E1122 09:25:23.322266 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:26 crc kubenswrapper[4735]: E1122 09:25:26.792937 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:28 crc kubenswrapper[4735]: I1122 09:25:28.984786 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-572lq" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" probeResult="failure" output=< Nov 22 09:25:28 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:25:28 crc kubenswrapper[4735]: > Nov 22 09:25:31 crc kubenswrapper[4735]: I1122 09:25:31.619602 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:31 crc kubenswrapper[4735]: I1122 09:25:31.696064 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:31 crc kubenswrapper[4735]: I1122 09:25:31.768673 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gvx69"] Nov 22 09:25:33 crc kubenswrapper[4735]: I1122 09:25:33.061632 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gvx69" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="registry-server" containerID="cri-o://1c51b907d875a6e5568da8f77da38f59223e824a21a8d04e68d66f0c3432f959" gracePeriod=2 Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.071893 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvx69" event={"ID":"0311e0ca-8871-4ca4-8123-138ec35497c5","Type":"ContainerDied","Data":"1c51b907d875a6e5568da8f77da38f59223e824a21a8d04e68d66f0c3432f959"} Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.073357 4735 generic.go:334] "Generic (PLEG): container finished" podID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerID="1c51b907d875a6e5568da8f77da38f59223e824a21a8d04e68d66f0c3432f959" exitCode=0 Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.525347 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.652647 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-catalog-content\") pod \"0311e0ca-8871-4ca4-8123-138ec35497c5\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.652999 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmvhs\" (UniqueName: \"kubernetes.io/projected/0311e0ca-8871-4ca4-8123-138ec35497c5-kube-api-access-cmvhs\") pod \"0311e0ca-8871-4ca4-8123-138ec35497c5\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.653099 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-utilities\") pod \"0311e0ca-8871-4ca4-8123-138ec35497c5\" (UID: \"0311e0ca-8871-4ca4-8123-138ec35497c5\") " Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.658630 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-utilities" (OuterVolumeSpecName: "utilities") pod "0311e0ca-8871-4ca4-8123-138ec35497c5" (UID: "0311e0ca-8871-4ca4-8123-138ec35497c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.703108 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0311e0ca-8871-4ca4-8123-138ec35497c5-kube-api-access-cmvhs" (OuterVolumeSpecName: "kube-api-access-cmvhs") pod "0311e0ca-8871-4ca4-8123-138ec35497c5" (UID: "0311e0ca-8871-4ca4-8123-138ec35497c5"). InnerVolumeSpecName "kube-api-access-cmvhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.759545 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmvhs\" (UniqueName: \"kubernetes.io/projected/0311e0ca-8871-4ca4-8123-138ec35497c5-kube-api-access-cmvhs\") on node \"crc\" DevicePath \"\"" Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.759586 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.845687 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0311e0ca-8871-4ca4-8123-138ec35497c5" (UID: "0311e0ca-8871-4ca4-8123-138ec35497c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:25:34 crc kubenswrapper[4735]: I1122 09:25:34.861387 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311e0ca-8871-4ca4-8123-138ec35497c5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:25:35 crc kubenswrapper[4735]: I1122 09:25:35.105286 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvx69" event={"ID":"0311e0ca-8871-4ca4-8123-138ec35497c5","Type":"ContainerDied","Data":"c3fdc699d31fea8f8f96c647ac584a8ce64b0b1ed4e6f4f83f015637d4dbcb3f"} Nov 22 09:25:35 crc kubenswrapper[4735]: I1122 09:25:35.105504 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvx69" Nov 22 09:25:35 crc kubenswrapper[4735]: I1122 09:25:35.110063 4735 scope.go:117] "RemoveContainer" containerID="1c51b907d875a6e5568da8f77da38f59223e824a21a8d04e68d66f0c3432f959" Nov 22 09:25:35 crc kubenswrapper[4735]: I1122 09:25:35.147816 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gvx69"] Nov 22 09:25:35 crc kubenswrapper[4735]: I1122 09:25:35.163182 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gvx69"] Nov 22 09:25:35 crc kubenswrapper[4735]: I1122 09:25:35.181047 4735 scope.go:117] "RemoveContainer" containerID="797b0cc556e53d4a14a5a0c2facb4808cec92e4b48a95485d4c45ce8040c55cc" Nov 22 09:25:35 crc kubenswrapper[4735]: I1122 09:25:35.210143 4735 scope.go:117] "RemoveContainer" containerID="cf7c4fad2b8bacd7345143f0aa131cd37b69377e7053e6fe42bd86b92020d087" Nov 22 09:25:35 crc kubenswrapper[4735]: I1122 09:25:35.283849 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" path="/var/lib/kubelet/pods/0311e0ca-8871-4ca4-8123-138ec35497c5/volumes" Nov 22 09:25:37 crc kubenswrapper[4735]: E1122 09:25:37.397682 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:38 crc kubenswrapper[4735]: E1122 09:25:38.080256 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:39 crc kubenswrapper[4735]: I1122 09:25:39.010001 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-572lq" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" probeResult="failure" output=< Nov 22 09:25:39 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:25:39 crc kubenswrapper[4735]: > Nov 22 09:25:46 crc kubenswrapper[4735]: I1122 09:25:46.132082 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:25:46 crc kubenswrapper[4735]: I1122 09:25:46.133573 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:25:47 crc kubenswrapper[4735]: E1122 09:25:47.980209 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:48 crc kubenswrapper[4735]: E1122 09:25:48.105961 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:48 crc kubenswrapper[4735]: E1122 09:25:48.111139 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:48 crc kubenswrapper[4735]: I1122 09:25:48.969549 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-572lq" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" probeResult="failure" output=< Nov 22 09:25:48 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:25:48 crc kubenswrapper[4735]: > Nov 22 09:25:53 crc kubenswrapper[4735]: E1122 09:25:53.263190 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:58 crc kubenswrapper[4735]: E1122 09:25:58.509565 4735 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0caf4ba5_8b95_4fa5_b479_5ed810e0d4ba.slice/crio-conmon-5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:25:59 crc kubenswrapper[4735]: I1122 09:25:59.303207 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-572lq" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" probeResult="failure" output=< Nov 22 09:25:59 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:25:59 crc kubenswrapper[4735]: > Nov 22 09:26:08 crc kubenswrapper[4735]: I1122 09:26:08.977826 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-572lq" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" probeResult="failure" output=< Nov 22 09:26:08 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:26:08 crc kubenswrapper[4735]: > Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.132058 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.136121 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.139874 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.146167 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.147383 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" gracePeriod=600 Nov 22 09:26:16 crc kubenswrapper[4735]: E1122 09:26:16.315159 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.671009 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" exitCode=0 Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.672533 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7"} Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.679052 4735 scope.go:117] "RemoveContainer" containerID="18f3c56d6b82827ef0d9798d93b1e5c6a5288d3b09223b912c74cba8a8ae1f73" Nov 22 09:26:16 crc kubenswrapper[4735]: I1122 09:26:16.679589 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:26:16 crc kubenswrapper[4735]: E1122 09:26:16.680930 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:26:18 crc kubenswrapper[4735]: I1122 09:26:18.010725 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:26:18 crc kubenswrapper[4735]: I1122 09:26:18.064489 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:26:18 crc kubenswrapper[4735]: I1122 09:26:18.167864 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-572lq"] Nov 22 09:26:19 crc kubenswrapper[4735]: I1122 09:26:19.713104 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-572lq" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" containerID="cri-o://c8d02a598743ac330c998d1451ffd32de338ea9dcbe283271f6bb0ddc9ac04e4" gracePeriod=2 Nov 22 09:26:20 crc kubenswrapper[4735]: I1122 09:26:20.756134 4735 generic.go:334] "Generic (PLEG): container finished" podID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerID="c8d02a598743ac330c998d1451ffd32de338ea9dcbe283271f6bb0ddc9ac04e4" exitCode=0 Nov 22 09:26:20 crc kubenswrapper[4735]: I1122 09:26:20.756319 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-572lq" event={"ID":"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba","Type":"ContainerDied","Data":"c8d02a598743ac330c998d1451ffd32de338ea9dcbe283271f6bb0ddc9ac04e4"} Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.088710 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.160805 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-utilities\") pod \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.160940 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7pnf\" (UniqueName: \"kubernetes.io/projected/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-kube-api-access-j7pnf\") pod \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.161049 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-catalog-content\") pod \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\" (UID: \"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba\") " Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.165875 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-utilities" (OuterVolumeSpecName: "utilities") pod "0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" (UID: "0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.192015 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-kube-api-access-j7pnf" (OuterVolumeSpecName: "kube-api-access-j7pnf") pod "0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" (UID: "0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba"). InnerVolumeSpecName "kube-api-access-j7pnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.264635 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.264669 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7pnf\" (UniqueName: \"kubernetes.io/projected/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-kube-api-access-j7pnf\") on node \"crc\" DevicePath \"\"" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.279451 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" (UID: "0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.366934 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.772335 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-572lq" event={"ID":"0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba","Type":"ContainerDied","Data":"8c1beefcd652f8f577c6af9c9da9912518329ba9c5993bc74b064cea05898557"} Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.772438 4735 scope.go:117] "RemoveContainer" containerID="c8d02a598743ac330c998d1451ffd32de338ea9dcbe283271f6bb0ddc9ac04e4" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.772507 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-572lq" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.812682 4735 scope.go:117] "RemoveContainer" containerID="32261606ce08528586b4beeabbbfe39eec3624f380b1addc7d9a5ba07611dd5b" Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.817432 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-572lq"] Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.829096 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-572lq"] Nov 22 09:26:21 crc kubenswrapper[4735]: I1122 09:26:21.843353 4735 scope.go:117] "RemoveContainer" containerID="5e1a7ce1c0f0e5b9299c1b5120b10469c32ba8a1dc465f3c54b5a58fb33b50b9" Nov 22 09:26:23 crc kubenswrapper[4735]: I1122 09:26:23.279279 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" path="/var/lib/kubelet/pods/0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba/volumes" Nov 22 09:26:29 crc kubenswrapper[4735]: I1122 09:26:29.271794 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:26:29 crc kubenswrapper[4735]: E1122 09:26:29.272446 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:26:41 crc kubenswrapper[4735]: I1122 09:26:41.265212 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:26:41 crc kubenswrapper[4735]: E1122 09:26:41.266628 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:26:55 crc kubenswrapper[4735]: I1122 09:26:55.264142 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:26:55 crc kubenswrapper[4735]: E1122 09:26:55.265210 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:27:09 crc kubenswrapper[4735]: I1122 09:27:09.282859 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:27:09 crc kubenswrapper[4735]: E1122 09:27:09.283850 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:27:22 crc kubenswrapper[4735]: I1122 09:27:22.262916 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:27:22 crc kubenswrapper[4735]: E1122 09:27:22.263596 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:27:36 crc kubenswrapper[4735]: I1122 09:27:36.263612 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:27:36 crc kubenswrapper[4735]: E1122 09:27:36.264660 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:27:50 crc kubenswrapper[4735]: I1122 09:27:50.264245 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:27:50 crc kubenswrapper[4735]: E1122 09:27:50.265266 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:28:04 crc kubenswrapper[4735]: I1122 09:28:04.264508 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:28:04 crc kubenswrapper[4735]: E1122 09:28:04.265476 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:28:16 crc kubenswrapper[4735]: I1122 09:28:16.263701 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:28:16 crc kubenswrapper[4735]: E1122 09:28:16.264689 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:28:29 crc kubenswrapper[4735]: I1122 09:28:29.280380 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:28:29 crc kubenswrapper[4735]: E1122 09:28:29.281589 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:28:41 crc kubenswrapper[4735]: I1122 09:28:41.265149 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:28:41 crc kubenswrapper[4735]: E1122 09:28:41.266031 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:28:52 crc kubenswrapper[4735]: I1122 09:28:52.263329 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:28:52 crc kubenswrapper[4735]: E1122 09:28:52.264146 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:29:06 crc kubenswrapper[4735]: I1122 09:29:06.264207 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:29:06 crc kubenswrapper[4735]: E1122 09:29:06.265018 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:29:20 crc kubenswrapper[4735]: I1122 09:29:20.264181 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:29:20 crc kubenswrapper[4735]: E1122 09:29:20.265018 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:29:31 crc kubenswrapper[4735]: I1122 09:29:31.267040 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:29:31 crc kubenswrapper[4735]: E1122 09:29:31.267942 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:29:45 crc kubenswrapper[4735]: I1122 09:29:45.264064 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:29:45 crc kubenswrapper[4735]: E1122 09:29:45.265392 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.121079 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nqzb5"] Nov 22 09:29:51 crc kubenswrapper[4735]: E1122 09:29:51.133409 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="extract-utilities" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.133472 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="extract-utilities" Nov 22 09:29:51 crc kubenswrapper[4735]: E1122 09:29:51.134155 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.134211 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" Nov 22 09:29:51 crc kubenswrapper[4735]: E1122 09:29:51.134280 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="extract-content" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.134302 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="extract-content" Nov 22 09:29:51 crc kubenswrapper[4735]: E1122 09:29:51.134338 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="extract-utilities" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.134348 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="extract-utilities" Nov 22 09:29:51 crc kubenswrapper[4735]: E1122 09:29:51.134366 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="registry-server" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.134374 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="registry-server" Nov 22 09:29:51 crc kubenswrapper[4735]: E1122 09:29:51.134398 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="extract-content" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.134407 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="extract-content" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.135853 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0311e0ca-8871-4ca4-8123-138ec35497c5" containerName="registry-server" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.135893 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="0caf4ba5-8b95-4fa5-b479-5ed810e0d4ba" containerName="registry-server" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.141249 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.212482 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqzb5"] Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.268817 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-utilities\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.269128 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-catalog-content\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.269218 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjvrm\" (UniqueName: \"kubernetes.io/projected/abbef4a9-b4d5-49aa-976e-081e6236b658-kube-api-access-bjvrm\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.371619 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-utilities\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.371711 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-catalog-content\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.371843 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjvrm\" (UniqueName: \"kubernetes.io/projected/abbef4a9-b4d5-49aa-976e-081e6236b658-kube-api-access-bjvrm\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.375319 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-catalog-content\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.375371 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-utilities\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.408385 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjvrm\" (UniqueName: \"kubernetes.io/projected/abbef4a9-b4d5-49aa-976e-081e6236b658-kube-api-access-bjvrm\") pod \"redhat-marketplace-nqzb5\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:51 crc kubenswrapper[4735]: I1122 09:29:51.477792 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:29:52 crc kubenswrapper[4735]: I1122 09:29:52.591567 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqzb5"] Nov 22 09:29:53 crc kubenswrapper[4735]: I1122 09:29:53.571751 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqzb5" event={"ID":"abbef4a9-b4d5-49aa-976e-081e6236b658","Type":"ContainerDied","Data":"bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28"} Nov 22 09:29:53 crc kubenswrapper[4735]: I1122 09:29:53.572169 4735 generic.go:334] "Generic (PLEG): container finished" podID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerID="bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28" exitCode=0 Nov 22 09:29:53 crc kubenswrapper[4735]: I1122 09:29:53.572485 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqzb5" event={"ID":"abbef4a9-b4d5-49aa-976e-081e6236b658","Type":"ContainerStarted","Data":"13c027ad35b4312446e98a66d98a1d9b1d81c852bc8da0f676df8d94d2f5f0b9"} Nov 22 09:29:53 crc kubenswrapper[4735]: I1122 09:29:53.578995 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:29:55 crc kubenswrapper[4735]: I1122 09:29:55.615882 4735 generic.go:334] "Generic (PLEG): container finished" podID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerID="68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd" exitCode=0 Nov 22 09:29:55 crc kubenswrapper[4735]: I1122 09:29:55.616232 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqzb5" event={"ID":"abbef4a9-b4d5-49aa-976e-081e6236b658","Type":"ContainerDied","Data":"68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd"} Nov 22 09:29:56 crc kubenswrapper[4735]: I1122 09:29:56.632240 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqzb5" event={"ID":"abbef4a9-b4d5-49aa-976e-081e6236b658","Type":"ContainerStarted","Data":"2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad"} Nov 22 09:29:56 crc kubenswrapper[4735]: I1122 09:29:56.663102 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nqzb5" podStartSLOduration=3.244437013 podStartE2EDuration="5.661748471s" podCreationTimestamp="2025-11-22 09:29:51 +0000 UTC" firstStartedPulling="2025-11-22 09:29:53.573697647 +0000 UTC m=+5215.178036242" lastFinishedPulling="2025-11-22 09:29:55.991009085 +0000 UTC m=+5217.595347700" observedRunningTime="2025-11-22 09:29:56.654656366 +0000 UTC m=+5218.258994981" watchObservedRunningTime="2025-11-22 09:29:56.661748471 +0000 UTC m=+5218.266087086" Nov 22 09:29:58 crc kubenswrapper[4735]: I1122 09:29:58.264426 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:29:58 crc kubenswrapper[4735]: E1122 09:29:58.266082 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.194141 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b"] Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.197245 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.207447 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.207755 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.210623 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b"] Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.296277 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/910bfab2-01a2-44be-8ba0-3ef696cd49eb-secret-volume\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.296749 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbk4b\" (UniqueName: \"kubernetes.io/projected/910bfab2-01a2-44be-8ba0-3ef696cd49eb-kube-api-access-nbk4b\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.296853 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/910bfab2-01a2-44be-8ba0-3ef696cd49eb-config-volume\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.399277 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/910bfab2-01a2-44be-8ba0-3ef696cd49eb-secret-volume\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.399411 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbk4b\" (UniqueName: \"kubernetes.io/projected/910bfab2-01a2-44be-8ba0-3ef696cd49eb-kube-api-access-nbk4b\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.399614 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/910bfab2-01a2-44be-8ba0-3ef696cd49eb-config-volume\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.400341 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/910bfab2-01a2-44be-8ba0-3ef696cd49eb-config-volume\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.406371 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/910bfab2-01a2-44be-8ba0-3ef696cd49eb-secret-volume\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.420588 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbk4b\" (UniqueName: \"kubernetes.io/projected/910bfab2-01a2-44be-8ba0-3ef696cd49eb-kube-api-access-nbk4b\") pod \"collect-profiles-29396730-gcm5b\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:00 crc kubenswrapper[4735]: I1122 09:30:00.531326 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:01 crc kubenswrapper[4735]: I1122 09:30:01.009312 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b"] Nov 22 09:30:01 crc kubenswrapper[4735]: W1122 09:30:01.014084 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod910bfab2_01a2_44be_8ba0_3ef696cd49eb.slice/crio-4c72d689c0d478ddae761acffd9b6922c8a0e2b911368890f0c02ce88576df0b WatchSource:0}: Error finding container 4c72d689c0d478ddae761acffd9b6922c8a0e2b911368890f0c02ce88576df0b: Status 404 returned error can't find the container with id 4c72d689c0d478ddae761acffd9b6922c8a0e2b911368890f0c02ce88576df0b Nov 22 09:30:01 crc kubenswrapper[4735]: I1122 09:30:01.478987 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:30:01 crc kubenswrapper[4735]: I1122 09:30:01.479060 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:30:01 crc kubenswrapper[4735]: I1122 09:30:01.707254 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" event={"ID":"910bfab2-01a2-44be-8ba0-3ef696cd49eb","Type":"ContainerStarted","Data":"8ab5e2526f9e9ac05f04dcdd31c2daef3fc7f5283e3560a9179488b1bd82b5f8"} Nov 22 09:30:01 crc kubenswrapper[4735]: I1122 09:30:01.707309 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" event={"ID":"910bfab2-01a2-44be-8ba0-3ef696cd49eb","Type":"ContainerStarted","Data":"4c72d689c0d478ddae761acffd9b6922c8a0e2b911368890f0c02ce88576df0b"} Nov 22 09:30:01 crc kubenswrapper[4735]: I1122 09:30:01.724912 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" podStartSLOduration=1.724888833 podStartE2EDuration="1.724888833s" podCreationTimestamp="2025-11-22 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 09:30:01.723254738 +0000 UTC m=+5223.327593363" watchObservedRunningTime="2025-11-22 09:30:01.724888833 +0000 UTC m=+5223.329227448" Nov 22 09:30:02 crc kubenswrapper[4735]: I1122 09:30:02.132985 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:30:02 crc kubenswrapper[4735]: I1122 09:30:02.187434 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:30:02 crc kubenswrapper[4735]: I1122 09:30:02.396497 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqzb5"] Nov 22 09:30:02 crc kubenswrapper[4735]: I1122 09:30:02.719307 4735 generic.go:334] "Generic (PLEG): container finished" podID="910bfab2-01a2-44be-8ba0-3ef696cd49eb" containerID="8ab5e2526f9e9ac05f04dcdd31c2daef3fc7f5283e3560a9179488b1bd82b5f8" exitCode=0 Nov 22 09:30:02 crc kubenswrapper[4735]: I1122 09:30:02.719365 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" event={"ID":"910bfab2-01a2-44be-8ba0-3ef696cd49eb","Type":"ContainerDied","Data":"8ab5e2526f9e9ac05f04dcdd31c2daef3fc7f5283e3560a9179488b1bd82b5f8"} Nov 22 09:30:03 crc kubenswrapper[4735]: I1122 09:30:03.730837 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nqzb5" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerName="registry-server" containerID="cri-o://2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad" gracePeriod=2 Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.236373 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.388991 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.408861 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/910bfab2-01a2-44be-8ba0-3ef696cd49eb-secret-volume\") pod \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.408930 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbk4b\" (UniqueName: \"kubernetes.io/projected/910bfab2-01a2-44be-8ba0-3ef696cd49eb-kube-api-access-nbk4b\") pod \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.408973 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/910bfab2-01a2-44be-8ba0-3ef696cd49eb-config-volume\") pod \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\" (UID: \"910bfab2-01a2-44be-8ba0-3ef696cd49eb\") " Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.411369 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/910bfab2-01a2-44be-8ba0-3ef696cd49eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "910bfab2-01a2-44be-8ba0-3ef696cd49eb" (UID: "910bfab2-01a2-44be-8ba0-3ef696cd49eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.417362 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910bfab2-01a2-44be-8ba0-3ef696cd49eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "910bfab2-01a2-44be-8ba0-3ef696cd49eb" (UID: "910bfab2-01a2-44be-8ba0-3ef696cd49eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.417542 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/910bfab2-01a2-44be-8ba0-3ef696cd49eb-kube-api-access-nbk4b" (OuterVolumeSpecName: "kube-api-access-nbk4b") pod "910bfab2-01a2-44be-8ba0-3ef696cd49eb" (UID: "910bfab2-01a2-44be-8ba0-3ef696cd49eb"). InnerVolumeSpecName "kube-api-access-nbk4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.511973 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjvrm\" (UniqueName: \"kubernetes.io/projected/abbef4a9-b4d5-49aa-976e-081e6236b658-kube-api-access-bjvrm\") pod \"abbef4a9-b4d5-49aa-976e-081e6236b658\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.512029 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-catalog-content\") pod \"abbef4a9-b4d5-49aa-976e-081e6236b658\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.512064 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-utilities\") pod \"abbef4a9-b4d5-49aa-976e-081e6236b658\" (UID: \"abbef4a9-b4d5-49aa-976e-081e6236b658\") " Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.512907 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbk4b\" (UniqueName: \"kubernetes.io/projected/910bfab2-01a2-44be-8ba0-3ef696cd49eb-kube-api-access-nbk4b\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.512927 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/910bfab2-01a2-44be-8ba0-3ef696cd49eb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.512938 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/910bfab2-01a2-44be-8ba0-3ef696cd49eb-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.513968 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-utilities" (OuterVolumeSpecName: "utilities") pod "abbef4a9-b4d5-49aa-976e-081e6236b658" (UID: "abbef4a9-b4d5-49aa-976e-081e6236b658"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.515576 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abbef4a9-b4d5-49aa-976e-081e6236b658-kube-api-access-bjvrm" (OuterVolumeSpecName: "kube-api-access-bjvrm") pod "abbef4a9-b4d5-49aa-976e-081e6236b658" (UID: "abbef4a9-b4d5-49aa-976e-081e6236b658"). InnerVolumeSpecName "kube-api-access-bjvrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.531154 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abbef4a9-b4d5-49aa-976e-081e6236b658" (UID: "abbef4a9-b4d5-49aa-976e-081e6236b658"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.616608 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjvrm\" (UniqueName: \"kubernetes.io/projected/abbef4a9-b4d5-49aa-976e-081e6236b658-kube-api-access-bjvrm\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.616679 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.617649 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abbef4a9-b4d5-49aa-976e-081e6236b658-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.755668 4735 generic.go:334] "Generic (PLEG): container finished" podID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerID="2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad" exitCode=0 Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.755896 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqzb5" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.755936 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqzb5" event={"ID":"abbef4a9-b4d5-49aa-976e-081e6236b658","Type":"ContainerDied","Data":"2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad"} Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.756199 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqzb5" event={"ID":"abbef4a9-b4d5-49aa-976e-081e6236b658","Type":"ContainerDied","Data":"13c027ad35b4312446e98a66d98a1d9b1d81c852bc8da0f676df8d94d2f5f0b9"} Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.756226 4735 scope.go:117] "RemoveContainer" containerID="2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.758731 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" event={"ID":"910bfab2-01a2-44be-8ba0-3ef696cd49eb","Type":"ContainerDied","Data":"4c72d689c0d478ddae761acffd9b6922c8a0e2b911368890f0c02ce88576df0b"} Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.758786 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-gcm5b" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.759369 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c72d689c0d478ddae761acffd9b6922c8a0e2b911368890f0c02ce88576df0b" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.808441 4735 scope.go:117] "RemoveContainer" containerID="68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.817647 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqzb5"] Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.830704 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d"] Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.833944 4735 scope.go:117] "RemoveContainer" containerID="bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.838502 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-rng9d"] Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.849020 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqzb5"] Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.920505 4735 scope.go:117] "RemoveContainer" containerID="2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad" Nov 22 09:30:04 crc kubenswrapper[4735]: E1122 09:30:04.921656 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad\": container with ID starting with 2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad not found: ID does not exist" containerID="2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.921698 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad"} err="failed to get container status \"2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad\": rpc error: code = NotFound desc = could not find container \"2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad\": container with ID starting with 2efd5c62f1325c1e59708645522ecee29318d39fb77b673cc75565127680c1ad not found: ID does not exist" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.921723 4735 scope.go:117] "RemoveContainer" containerID="68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd" Nov 22 09:30:04 crc kubenswrapper[4735]: E1122 09:30:04.922215 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd\": container with ID starting with 68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd not found: ID does not exist" containerID="68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.922269 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd"} err="failed to get container status \"68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd\": rpc error: code = NotFound desc = could not find container \"68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd\": container with ID starting with 68ef28f9861a9e9129e6618df9d1d92c58e6d8af815bdba9e567bb6a78b1e7dd not found: ID does not exist" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.922305 4735 scope.go:117] "RemoveContainer" containerID="bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28" Nov 22 09:30:04 crc kubenswrapper[4735]: E1122 09:30:04.922843 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28\": container with ID starting with bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28 not found: ID does not exist" containerID="bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28" Nov 22 09:30:04 crc kubenswrapper[4735]: I1122 09:30:04.923040 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28"} err="failed to get container status \"bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28\": rpc error: code = NotFound desc = could not find container \"bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28\": container with ID starting with bec1ea83be5e81c7984f830873e684ac04288ede1954cb2f197361a44dd33f28 not found: ID does not exist" Nov 22 09:30:05 crc kubenswrapper[4735]: I1122 09:30:05.278891 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8" path="/var/lib/kubelet/pods/45a5c2f9-4791-4de3-8cc6-3859fdc4d2f8/volumes" Nov 22 09:30:05 crc kubenswrapper[4735]: I1122 09:30:05.280227 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" path="/var/lib/kubelet/pods/abbef4a9-b4d5-49aa-976e-081e6236b658/volumes" Nov 22 09:30:09 crc kubenswrapper[4735]: I1122 09:30:09.270571 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:30:09 crc kubenswrapper[4735]: E1122 09:30:09.271395 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:30:12 crc kubenswrapper[4735]: I1122 09:30:12.048029 4735 scope.go:117] "RemoveContainer" containerID="e71a621fbeea5d717029f1490c681f4e9cadffeefe63d3dba96a90050fccf2cb" Nov 22 09:30:23 crc kubenswrapper[4735]: I1122 09:30:23.263866 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:30:23 crc kubenswrapper[4735]: E1122 09:30:23.265090 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:30:36 crc kubenswrapper[4735]: I1122 09:30:36.263936 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:30:36 crc kubenswrapper[4735]: E1122 09:30:36.265128 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:30:51 crc kubenswrapper[4735]: I1122 09:30:51.263174 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:30:51 crc kubenswrapper[4735]: E1122 09:30:51.263818 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:31:03 crc kubenswrapper[4735]: I1122 09:31:03.263395 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:31:03 crc kubenswrapper[4735]: E1122 09:31:03.264203 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:31:16 crc kubenswrapper[4735]: I1122 09:31:16.264304 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:31:16 crc kubenswrapper[4735]: I1122 09:31:16.666738 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"ab9ddf2de4a6cb93d17c04b33b3cf8f35dcb10a57394d20f90dceb88ba6ec85c"} Nov 22 09:32:42 crc kubenswrapper[4735]: I1122 09:32:42.177941 4735 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.455721517s: [/var/lib/containers/storage/overlay/9086dd1d8bbc0adf4b5bd778ea4ee1de4ca909a065494f353a67227bd0e7372b/diff /var/log/pods/openstack_openstackclient_da8872ad-7c19-451d-84be-6333126bfb63/openstackclient/0.log]; will not log again for this container unless duration exceeds 2s Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.163756 4735 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 5.190818571s: [/var/lib/containers/storage/overlay/3fb8de1cd1d75ef6665703c51ae49f9464c7cc5b8d9afa5e8386262530380a0a/diff /var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-vgsgb_3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c/manager/0.log]; will not log again for this container unless duration exceeds 2s Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.165631 4735 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.278480007s: [/var/lib/containers/storage/overlay/71c10dc7d73e7bce0546d44be1f02432688c4d609f318090c65e83cceb36e83a/diff /var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-xl2mv_9124e032-f3c0-46fa-bd17-26ed40bfd57e/manager/0.log]; will not log again for this container unless duration exceeds 2s Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.168660 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="be212b2c-6753-40d9-9bee-0dfe2ff10dbc" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.170046 4735 trace.go:236] Trace[201396601]: "Calculate volume metrics of prometheus-metric-storage-rulefiles-0 for pod openstack/prometheus-metric-storage-0" (22-Nov-2025 09:32:47.582) (total time: 5582ms): Nov 22 09:32:53 crc kubenswrapper[4735]: Trace[201396601]: [5.582170899s] [5.582170899s] END Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.175065 4735 patch_prober.go:28] interesting pod/logging-loki-gateway-6f9c8b877-5dtcw container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.77:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.175119 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" podUID="5caa9595-315a-40a7-a46f-5e6d4b13c320" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.77:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.179173 4735 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-8vp7b container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.179243 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-8vp7b" podUID="5b742605-ede0-46c6-add0-01cda9fa184d" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 09:32:53 crc kubenswrapper[4735]: I1122 09:32:53.185644 4735 trace.go:236] Trace[1781250594]: "Calculate volume metrics of registry-certificates for pod openshift-image-registry/image-registry-66df7c8f76-wmm22" (22-Nov-2025 09:32:52.152) (total time: 1032ms): Nov 22 09:32:53 crc kubenswrapper[4735]: Trace[1781250594]: [1.032748943s] [1.032748943s] END Nov 22 09:33:16 crc kubenswrapper[4735]: I1122 09:33:16.131368 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:33:16 crc kubenswrapper[4735]: I1122 09:33:16.131814 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.702074 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rj2jp"] Nov 22 09:33:26 crc kubenswrapper[4735]: E1122 09:33:26.703200 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerName="extract-content" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.703217 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerName="extract-content" Nov 22 09:33:26 crc kubenswrapper[4735]: E1122 09:33:26.703261 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerName="registry-server" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.703270 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerName="registry-server" Nov 22 09:33:26 crc kubenswrapper[4735]: E1122 09:33:26.703312 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerName="extract-utilities" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.703320 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerName="extract-utilities" Nov 22 09:33:26 crc kubenswrapper[4735]: E1122 09:33:26.703334 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910bfab2-01a2-44be-8ba0-3ef696cd49eb" containerName="collect-profiles" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.703342 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="910bfab2-01a2-44be-8ba0-3ef696cd49eb" containerName="collect-profiles" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.703699 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="abbef4a9-b4d5-49aa-976e-081e6236b658" containerName="registry-server" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.703738 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="910bfab2-01a2-44be-8ba0-3ef696cd49eb" containerName="collect-profiles" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.705999 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.724193 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rj2jp"] Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.870398 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-utilities\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.870439 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-catalog-content\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.870517 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bqw8\" (UniqueName: \"kubernetes.io/projected/09a28b49-f4ce-4e79-897a-f0308e6b5691-kube-api-access-5bqw8\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.973381 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-utilities\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.973438 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-catalog-content\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.973552 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bqw8\" (UniqueName: \"kubernetes.io/projected/09a28b49-f4ce-4e79-897a-f0308e6b5691-kube-api-access-5bqw8\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.978757 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-utilities\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:26 crc kubenswrapper[4735]: I1122 09:33:26.980074 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-catalog-content\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:27 crc kubenswrapper[4735]: I1122 09:33:27.007830 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bqw8\" (UniqueName: \"kubernetes.io/projected/09a28b49-f4ce-4e79-897a-f0308e6b5691-kube-api-access-5bqw8\") pod \"certified-operators-rj2jp\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:27 crc kubenswrapper[4735]: I1122 09:33:27.040065 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:27 crc kubenswrapper[4735]: I1122 09:33:27.601779 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rj2jp"] Nov 22 09:33:27 crc kubenswrapper[4735]: W1122 09:33:27.621341 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09a28b49_f4ce_4e79_897a_f0308e6b5691.slice/crio-72898c6b814fa3221a03a182f08ce0c12d1125a8cbd5ac8373da335c5b406dd4 WatchSource:0}: Error finding container 72898c6b814fa3221a03a182f08ce0c12d1125a8cbd5ac8373da335c5b406dd4: Status 404 returned error can't find the container with id 72898c6b814fa3221a03a182f08ce0c12d1125a8cbd5ac8373da335c5b406dd4 Nov 22 09:33:27 crc kubenswrapper[4735]: I1122 09:33:27.655574 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj2jp" event={"ID":"09a28b49-f4ce-4e79-897a-f0308e6b5691","Type":"ContainerStarted","Data":"72898c6b814fa3221a03a182f08ce0c12d1125a8cbd5ac8373da335c5b406dd4"} Nov 22 09:33:28 crc kubenswrapper[4735]: I1122 09:33:28.670291 4735 generic.go:334] "Generic (PLEG): container finished" podID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerID="ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63" exitCode=0 Nov 22 09:33:28 crc kubenswrapper[4735]: I1122 09:33:28.670413 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj2jp" event={"ID":"09a28b49-f4ce-4e79-897a-f0308e6b5691","Type":"ContainerDied","Data":"ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63"} Nov 22 09:33:29 crc kubenswrapper[4735]: I1122 09:33:29.687955 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj2jp" event={"ID":"09a28b49-f4ce-4e79-897a-f0308e6b5691","Type":"ContainerStarted","Data":"d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6"} Nov 22 09:33:31 crc kubenswrapper[4735]: I1122 09:33:31.717770 4735 generic.go:334] "Generic (PLEG): container finished" podID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerID="d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6" exitCode=0 Nov 22 09:33:31 crc kubenswrapper[4735]: I1122 09:33:31.717847 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj2jp" event={"ID":"09a28b49-f4ce-4e79-897a-f0308e6b5691","Type":"ContainerDied","Data":"d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6"} Nov 22 09:33:32 crc kubenswrapper[4735]: I1122 09:33:32.735556 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj2jp" event={"ID":"09a28b49-f4ce-4e79-897a-f0308e6b5691","Type":"ContainerStarted","Data":"5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a"} Nov 22 09:33:32 crc kubenswrapper[4735]: I1122 09:33:32.769086 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rj2jp" podStartSLOduration=3.318479522 podStartE2EDuration="6.769066884s" podCreationTimestamp="2025-11-22 09:33:26 +0000 UTC" firstStartedPulling="2025-11-22 09:33:28.673080326 +0000 UTC m=+5430.277418931" lastFinishedPulling="2025-11-22 09:33:32.123667678 +0000 UTC m=+5433.728006293" observedRunningTime="2025-11-22 09:33:32.765659 +0000 UTC m=+5434.369997605" watchObservedRunningTime="2025-11-22 09:33:32.769066884 +0000 UTC m=+5434.373405489" Nov 22 09:33:37 crc kubenswrapper[4735]: I1122 09:33:37.041150 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:37 crc kubenswrapper[4735]: I1122 09:33:37.041756 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:38 crc kubenswrapper[4735]: I1122 09:33:38.090535 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rj2jp" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="registry-server" probeResult="failure" output=< Nov 22 09:33:38 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:33:38 crc kubenswrapper[4735]: > Nov 22 09:33:46 crc kubenswrapper[4735]: I1122 09:33:46.131580 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:33:46 crc kubenswrapper[4735]: I1122 09:33:46.132105 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:33:47 crc kubenswrapper[4735]: I1122 09:33:47.109974 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:47 crc kubenswrapper[4735]: I1122 09:33:47.199348 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:47 crc kubenswrapper[4735]: I1122 09:33:47.353316 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rj2jp"] Nov 22 09:33:48 crc kubenswrapper[4735]: I1122 09:33:48.902338 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rj2jp" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="registry-server" containerID="cri-o://5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a" gracePeriod=2 Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.459352 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.536132 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-utilities\") pod \"09a28b49-f4ce-4e79-897a-f0308e6b5691\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.536229 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-catalog-content\") pod \"09a28b49-f4ce-4e79-897a-f0308e6b5691\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.536386 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bqw8\" (UniqueName: \"kubernetes.io/projected/09a28b49-f4ce-4e79-897a-f0308e6b5691-kube-api-access-5bqw8\") pod \"09a28b49-f4ce-4e79-897a-f0308e6b5691\" (UID: \"09a28b49-f4ce-4e79-897a-f0308e6b5691\") " Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.543852 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09a28b49-f4ce-4e79-897a-f0308e6b5691-kube-api-access-5bqw8" (OuterVolumeSpecName: "kube-api-access-5bqw8") pod "09a28b49-f4ce-4e79-897a-f0308e6b5691" (UID: "09a28b49-f4ce-4e79-897a-f0308e6b5691"). InnerVolumeSpecName "kube-api-access-5bqw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.544355 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-utilities" (OuterVolumeSpecName: "utilities") pod "09a28b49-f4ce-4e79-897a-f0308e6b5691" (UID: "09a28b49-f4ce-4e79-897a-f0308e6b5691"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.584009 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09a28b49-f4ce-4e79-897a-f0308e6b5691" (UID: "09a28b49-f4ce-4e79-897a-f0308e6b5691"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.639171 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.639205 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a28b49-f4ce-4e79-897a-f0308e6b5691-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.639217 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bqw8\" (UniqueName: \"kubernetes.io/projected/09a28b49-f4ce-4e79-897a-f0308e6b5691-kube-api-access-5bqw8\") on node \"crc\" DevicePath \"\"" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.920039 4735 generic.go:334] "Generic (PLEG): container finished" podID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerID="5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a" exitCode=0 Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.920091 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj2jp" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.920104 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj2jp" event={"ID":"09a28b49-f4ce-4e79-897a-f0308e6b5691","Type":"ContainerDied","Data":"5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a"} Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.920151 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj2jp" event={"ID":"09a28b49-f4ce-4e79-897a-f0308e6b5691","Type":"ContainerDied","Data":"72898c6b814fa3221a03a182f08ce0c12d1125a8cbd5ac8373da335c5b406dd4"} Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.920180 4735 scope.go:117] "RemoveContainer" containerID="5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.947328 4735 scope.go:117] "RemoveContainer" containerID="d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.979508 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rj2jp"] Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.989480 4735 scope.go:117] "RemoveContainer" containerID="ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63" Nov 22 09:33:49 crc kubenswrapper[4735]: I1122 09:33:49.997274 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rj2jp"] Nov 22 09:33:50 crc kubenswrapper[4735]: I1122 09:33:50.038165 4735 scope.go:117] "RemoveContainer" containerID="5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a" Nov 22 09:33:50 crc kubenswrapper[4735]: E1122 09:33:50.038747 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a\": container with ID starting with 5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a not found: ID does not exist" containerID="5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a" Nov 22 09:33:50 crc kubenswrapper[4735]: I1122 09:33:50.038783 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a"} err="failed to get container status \"5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a\": rpc error: code = NotFound desc = could not find container \"5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a\": container with ID starting with 5adfd8f22da967f4d2c84c92e17f230377f6d67b4c5d4dc94bb53b53bada468a not found: ID does not exist" Nov 22 09:33:50 crc kubenswrapper[4735]: I1122 09:33:50.038807 4735 scope.go:117] "RemoveContainer" containerID="d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6" Nov 22 09:33:50 crc kubenswrapper[4735]: E1122 09:33:50.039194 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6\": container with ID starting with d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6 not found: ID does not exist" containerID="d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6" Nov 22 09:33:50 crc kubenswrapper[4735]: I1122 09:33:50.039234 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6"} err="failed to get container status \"d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6\": rpc error: code = NotFound desc = could not find container \"d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6\": container with ID starting with d852e0f5ca6f3e915bae67e8b25e2587833d915737b19428c7672a16764631b6 not found: ID does not exist" Nov 22 09:33:50 crc kubenswrapper[4735]: I1122 09:33:50.039253 4735 scope.go:117] "RemoveContainer" containerID="ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63" Nov 22 09:33:50 crc kubenswrapper[4735]: E1122 09:33:50.039679 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63\": container with ID starting with ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63 not found: ID does not exist" containerID="ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63" Nov 22 09:33:50 crc kubenswrapper[4735]: I1122 09:33:50.039820 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63"} err="failed to get container status \"ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63\": rpc error: code = NotFound desc = could not find container \"ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63\": container with ID starting with ba4467387f7d7b85d3fd7b06107423bacf979dae31989e6eb10c38dc53b0cf63 not found: ID does not exist" Nov 22 09:33:51 crc kubenswrapper[4735]: I1122 09:33:51.281726 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" path="/var/lib/kubelet/pods/09a28b49-f4ce-4e79-897a-f0308e6b5691/volumes" Nov 22 09:34:16 crc kubenswrapper[4735]: I1122 09:34:16.131962 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:34:16 crc kubenswrapper[4735]: I1122 09:34:16.132610 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:34:16 crc kubenswrapper[4735]: I1122 09:34:16.132699 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:34:16 crc kubenswrapper[4735]: I1122 09:34:16.133610 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab9ddf2de4a6cb93d17c04b33b3cf8f35dcb10a57394d20f90dceb88ba6ec85c"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:34:16 crc kubenswrapper[4735]: I1122 09:34:16.133658 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://ab9ddf2de4a6cb93d17c04b33b3cf8f35dcb10a57394d20f90dceb88ba6ec85c" gracePeriod=600 Nov 22 09:34:16 crc kubenswrapper[4735]: I1122 09:34:16.303079 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="ab9ddf2de4a6cb93d17c04b33b3cf8f35dcb10a57394d20f90dceb88ba6ec85c" exitCode=0 Nov 22 09:34:16 crc kubenswrapper[4735]: I1122 09:34:16.303153 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"ab9ddf2de4a6cb93d17c04b33b3cf8f35dcb10a57394d20f90dceb88ba6ec85c"} Nov 22 09:34:16 crc kubenswrapper[4735]: I1122 09:34:16.303513 4735 scope.go:117] "RemoveContainer" containerID="13d125018d40e0ef794dbb9ba46ebf09eb2b87e5206e79fc606eba0bbbd201b7" Nov 22 09:34:17 crc kubenswrapper[4735]: I1122 09:34:17.320617 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4"} Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.282006 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k2t6s"] Nov 22 09:35:08 crc kubenswrapper[4735]: E1122 09:35:08.282964 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="extract-content" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.282977 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="extract-content" Nov 22 09:35:08 crc kubenswrapper[4735]: E1122 09:35:08.282991 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="extract-utilities" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.282997 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="extract-utilities" Nov 22 09:35:08 crc kubenswrapper[4735]: E1122 09:35:08.283018 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="registry-server" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.283023 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="registry-server" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.285596 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="09a28b49-f4ce-4e79-897a-f0308e6b5691" containerName="registry-server" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.287864 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.319865 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2t6s"] Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.405753 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-catalog-content\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.405891 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-utilities\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.406015 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7667\" (UniqueName: \"kubernetes.io/projected/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-kube-api-access-r7667\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.508702 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-catalog-content\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.508899 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-utilities\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.509170 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7667\" (UniqueName: \"kubernetes.io/projected/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-kube-api-access-r7667\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.509394 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-catalog-content\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.509417 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-utilities\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.530835 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7667\" (UniqueName: \"kubernetes.io/projected/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-kube-api-access-r7667\") pod \"community-operators-k2t6s\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:08 crc kubenswrapper[4735]: I1122 09:35:08.627129 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:09 crc kubenswrapper[4735]: I1122 09:35:09.428882 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2t6s"] Nov 22 09:35:10 crc kubenswrapper[4735]: W1122 09:35:10.064766 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ee3f179_5bf3_46a6_8980_0acfea0cd1cb.slice/crio-95611d416c946f6da844c23b543c52c8a94693f593f3452017a8c71d60b445a1 WatchSource:0}: Error finding container 95611d416c946f6da844c23b543c52c8a94693f593f3452017a8c71d60b445a1: Status 404 returned error can't find the container with id 95611d416c946f6da844c23b543c52c8a94693f593f3452017a8c71d60b445a1 Nov 22 09:35:10 crc kubenswrapper[4735]: I1122 09:35:10.996034 4735 generic.go:334] "Generic (PLEG): container finished" podID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerID="45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34" exitCode=0 Nov 22 09:35:10 crc kubenswrapper[4735]: I1122 09:35:10.996181 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2t6s" event={"ID":"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb","Type":"ContainerDied","Data":"45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34"} Nov 22 09:35:10 crc kubenswrapper[4735]: I1122 09:35:10.996368 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2t6s" event={"ID":"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb","Type":"ContainerStarted","Data":"95611d416c946f6da844c23b543c52c8a94693f593f3452017a8c71d60b445a1"} Nov 22 09:35:10 crc kubenswrapper[4735]: I1122 09:35:10.998947 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:35:12 crc kubenswrapper[4735]: I1122 09:35:12.892235 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l9m58"] Nov 22 09:35:12 crc kubenswrapper[4735]: I1122 09:35:12.895280 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:12 crc kubenswrapper[4735]: I1122 09:35:12.917240 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2rtg\" (UniqueName: \"kubernetes.io/projected/a4f9a546-9052-4d89-8997-e826c9d8ed75-kube-api-access-v2rtg\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:12 crc kubenswrapper[4735]: I1122 09:35:12.917476 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-catalog-content\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:12 crc kubenswrapper[4735]: I1122 09:35:12.917632 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-utilities\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:12 crc kubenswrapper[4735]: I1122 09:35:12.923572 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9m58"] Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.019267 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-utilities\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.019599 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2rtg\" (UniqueName: \"kubernetes.io/projected/a4f9a546-9052-4d89-8997-e826c9d8ed75-kube-api-access-v2rtg\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.019746 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-catalog-content\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.019860 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-utilities\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.020124 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-catalog-content\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.026364 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2t6s" event={"ID":"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb","Type":"ContainerStarted","Data":"888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b"} Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.043827 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2rtg\" (UniqueName: \"kubernetes.io/projected/a4f9a546-9052-4d89-8997-e826c9d8ed75-kube-api-access-v2rtg\") pod \"redhat-operators-l9m58\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.228658 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:13 crc kubenswrapper[4735]: I1122 09:35:13.806306 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9m58"] Nov 22 09:35:14 crc kubenswrapper[4735]: I1122 09:35:14.037519 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9m58" event={"ID":"a4f9a546-9052-4d89-8997-e826c9d8ed75","Type":"ContainerStarted","Data":"4ac6efbf07f116e7a184c603b3aa7b78d2abb04076360b7dfd3093242e8b994a"} Nov 22 09:35:15 crc kubenswrapper[4735]: I1122 09:35:15.056673 4735 generic.go:334] "Generic (PLEG): container finished" podID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerID="f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3" exitCode=0 Nov 22 09:35:15 crc kubenswrapper[4735]: I1122 09:35:15.056785 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9m58" event={"ID":"a4f9a546-9052-4d89-8997-e826c9d8ed75","Type":"ContainerDied","Data":"f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3"} Nov 22 09:35:15 crc kubenswrapper[4735]: I1122 09:35:15.061204 4735 generic.go:334] "Generic (PLEG): container finished" podID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerID="888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b" exitCode=0 Nov 22 09:35:15 crc kubenswrapper[4735]: I1122 09:35:15.061254 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2t6s" event={"ID":"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb","Type":"ContainerDied","Data":"888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b"} Nov 22 09:35:16 crc kubenswrapper[4735]: I1122 09:35:16.080066 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2t6s" event={"ID":"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb","Type":"ContainerStarted","Data":"a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386"} Nov 22 09:35:16 crc kubenswrapper[4735]: I1122 09:35:16.114150 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k2t6s" podStartSLOduration=3.581220053 podStartE2EDuration="8.114129899s" podCreationTimestamp="2025-11-22 09:35:08 +0000 UTC" firstStartedPulling="2025-11-22 09:35:10.998516614 +0000 UTC m=+5532.602855269" lastFinishedPulling="2025-11-22 09:35:15.53142647 +0000 UTC m=+5537.135765115" observedRunningTime="2025-11-22 09:35:16.105085832 +0000 UTC m=+5537.709424447" watchObservedRunningTime="2025-11-22 09:35:16.114129899 +0000 UTC m=+5537.718468504" Nov 22 09:35:17 crc kubenswrapper[4735]: I1122 09:35:17.103651 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9m58" event={"ID":"a4f9a546-9052-4d89-8997-e826c9d8ed75","Type":"ContainerStarted","Data":"ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55"} Nov 22 09:35:18 crc kubenswrapper[4735]: I1122 09:35:18.627925 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:18 crc kubenswrapper[4735]: I1122 09:35:18.628273 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:19 crc kubenswrapper[4735]: I1122 09:35:19.854773 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-k2t6s" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="registry-server" probeResult="failure" output=< Nov 22 09:35:19 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:35:19 crc kubenswrapper[4735]: > Nov 22 09:35:24 crc kubenswrapper[4735]: I1122 09:35:24.205355 4735 generic.go:334] "Generic (PLEG): container finished" podID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerID="ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55" exitCode=0 Nov 22 09:35:24 crc kubenswrapper[4735]: I1122 09:35:24.205429 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9m58" event={"ID":"a4f9a546-9052-4d89-8997-e826c9d8ed75","Type":"ContainerDied","Data":"ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55"} Nov 22 09:35:26 crc kubenswrapper[4735]: I1122 09:35:26.230409 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9m58" event={"ID":"a4f9a546-9052-4d89-8997-e826c9d8ed75","Type":"ContainerStarted","Data":"e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d"} Nov 22 09:35:26 crc kubenswrapper[4735]: I1122 09:35:26.266064 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l9m58" podStartSLOduration=4.674718134 podStartE2EDuration="14.266043172s" podCreationTimestamp="2025-11-22 09:35:12 +0000 UTC" firstStartedPulling="2025-11-22 09:35:15.06070208 +0000 UTC m=+5536.665040725" lastFinishedPulling="2025-11-22 09:35:24.652027118 +0000 UTC m=+5546.256365763" observedRunningTime="2025-11-22 09:35:26.256029768 +0000 UTC m=+5547.860368383" watchObservedRunningTime="2025-11-22 09:35:26.266043172 +0000 UTC m=+5547.870381787" Nov 22 09:35:29 crc kubenswrapper[4735]: I1122 09:35:29.688860 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-k2t6s" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="registry-server" probeResult="failure" output=< Nov 22 09:35:29 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:35:29 crc kubenswrapper[4735]: > Nov 22 09:35:33 crc kubenswrapper[4735]: I1122 09:35:33.228864 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:33 crc kubenswrapper[4735]: I1122 09:35:33.230780 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:34 crc kubenswrapper[4735]: I1122 09:35:34.281962 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9m58" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="registry-server" probeResult="failure" output=< Nov 22 09:35:34 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:35:34 crc kubenswrapper[4735]: > Nov 22 09:35:38 crc kubenswrapper[4735]: I1122 09:35:38.708436 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:38 crc kubenswrapper[4735]: I1122 09:35:38.777842 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:39 crc kubenswrapper[4735]: I1122 09:35:39.069976 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2t6s"] Nov 22 09:35:40 crc kubenswrapper[4735]: I1122 09:35:40.381003 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k2t6s" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="registry-server" containerID="cri-o://a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386" gracePeriod=2 Nov 22 09:35:40 crc kubenswrapper[4735]: I1122 09:35:40.957239 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.122385 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-catalog-content\") pod \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.122788 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7667\" (UniqueName: \"kubernetes.io/projected/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-kube-api-access-r7667\") pod \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.122923 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-utilities\") pod \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\" (UID: \"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb\") " Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.125168 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-utilities" (OuterVolumeSpecName: "utilities") pod "8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" (UID: "8ee3f179-5bf3-46a6-8980-0acfea0cd1cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.131665 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-kube-api-access-r7667" (OuterVolumeSpecName: "kube-api-access-r7667") pod "8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" (UID: "8ee3f179-5bf3-46a6-8980-0acfea0cd1cb"). InnerVolumeSpecName "kube-api-access-r7667". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.195474 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" (UID: "8ee3f179-5bf3-46a6-8980-0acfea0cd1cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.225320 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7667\" (UniqueName: \"kubernetes.io/projected/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-kube-api-access-r7667\") on node \"crc\" DevicePath \"\"" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.225346 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.225373 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.396693 4735 generic.go:334] "Generic (PLEG): container finished" podID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerID="a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386" exitCode=0 Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.396745 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2t6s" event={"ID":"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb","Type":"ContainerDied","Data":"a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386"} Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.396757 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2t6s" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.396771 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2t6s" event={"ID":"8ee3f179-5bf3-46a6-8980-0acfea0cd1cb","Type":"ContainerDied","Data":"95611d416c946f6da844c23b543c52c8a94693f593f3452017a8c71d60b445a1"} Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.396789 4735 scope.go:117] "RemoveContainer" containerID="a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.435237 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2t6s"] Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.444624 4735 scope.go:117] "RemoveContainer" containerID="888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.449599 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k2t6s"] Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.477452 4735 scope.go:117] "RemoveContainer" containerID="45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.536514 4735 scope.go:117] "RemoveContainer" containerID="a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386" Nov 22 09:35:41 crc kubenswrapper[4735]: E1122 09:35:41.537008 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386\": container with ID starting with a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386 not found: ID does not exist" containerID="a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.537064 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386"} err="failed to get container status \"a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386\": rpc error: code = NotFound desc = could not find container \"a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386\": container with ID starting with a0ea965da14e539109309c77a9c3866b47d49ad345989d26236e2eee5e3ee386 not found: ID does not exist" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.537097 4735 scope.go:117] "RemoveContainer" containerID="888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b" Nov 22 09:35:41 crc kubenswrapper[4735]: E1122 09:35:41.537399 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b\": container with ID starting with 888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b not found: ID does not exist" containerID="888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.537423 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b"} err="failed to get container status \"888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b\": rpc error: code = NotFound desc = could not find container \"888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b\": container with ID starting with 888f409a9a84997378a0be672d166901fc09b1ff17d7e87751b0dc8449f7f46b not found: ID does not exist" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.537436 4735 scope.go:117] "RemoveContainer" containerID="45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34" Nov 22 09:35:41 crc kubenswrapper[4735]: E1122 09:35:41.538075 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34\": container with ID starting with 45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34 not found: ID does not exist" containerID="45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34" Nov 22 09:35:41 crc kubenswrapper[4735]: I1122 09:35:41.538123 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34"} err="failed to get container status \"45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34\": rpc error: code = NotFound desc = could not find container \"45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34\": container with ID starting with 45b30eef4ce0988a968b32f83d99aa2588a995dc82c8f27e059e1eb7c378ce34 not found: ID does not exist" Nov 22 09:35:43 crc kubenswrapper[4735]: I1122 09:35:43.280512 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" path="/var/lib/kubelet/pods/8ee3f179-5bf3-46a6-8980-0acfea0cd1cb/volumes" Nov 22 09:35:44 crc kubenswrapper[4735]: I1122 09:35:44.299573 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9m58" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="registry-server" probeResult="failure" output=< Nov 22 09:35:44 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:35:44 crc kubenswrapper[4735]: > Nov 22 09:35:53 crc kubenswrapper[4735]: I1122 09:35:53.292421 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:53 crc kubenswrapper[4735]: I1122 09:35:53.365956 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:53 crc kubenswrapper[4735]: I1122 09:35:53.536225 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9m58"] Nov 22 09:35:54 crc kubenswrapper[4735]: I1122 09:35:54.533427 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l9m58" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="registry-server" containerID="cri-o://e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d" gracePeriod=2 Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.370875 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.478635 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-catalog-content\") pod \"a4f9a546-9052-4d89-8997-e826c9d8ed75\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.478812 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2rtg\" (UniqueName: \"kubernetes.io/projected/a4f9a546-9052-4d89-8997-e826c9d8ed75-kube-api-access-v2rtg\") pod \"a4f9a546-9052-4d89-8997-e826c9d8ed75\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.479099 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-utilities\") pod \"a4f9a546-9052-4d89-8997-e826c9d8ed75\" (UID: \"a4f9a546-9052-4d89-8997-e826c9d8ed75\") " Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.480690 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-utilities" (OuterVolumeSpecName: "utilities") pod "a4f9a546-9052-4d89-8997-e826c9d8ed75" (UID: "a4f9a546-9052-4d89-8997-e826c9d8ed75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.485990 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4f9a546-9052-4d89-8997-e826c9d8ed75-kube-api-access-v2rtg" (OuterVolumeSpecName: "kube-api-access-v2rtg") pod "a4f9a546-9052-4d89-8997-e826c9d8ed75" (UID: "a4f9a546-9052-4d89-8997-e826c9d8ed75"). InnerVolumeSpecName "kube-api-access-v2rtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.546203 4735 generic.go:334] "Generic (PLEG): container finished" podID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerID="e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d" exitCode=0 Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.546407 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9m58" event={"ID":"a4f9a546-9052-4d89-8997-e826c9d8ed75","Type":"ContainerDied","Data":"e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d"} Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.547377 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9m58" event={"ID":"a4f9a546-9052-4d89-8997-e826c9d8ed75","Type":"ContainerDied","Data":"4ac6efbf07f116e7a184c603b3aa7b78d2abb04076360b7dfd3093242e8b994a"} Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.546568 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9m58" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.547519 4735 scope.go:117] "RemoveContainer" containerID="e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.574738 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4f9a546-9052-4d89-8997-e826c9d8ed75" (UID: "a4f9a546-9052-4d89-8997-e826c9d8ed75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.578704 4735 scope.go:117] "RemoveContainer" containerID="ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.582800 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.582832 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2rtg\" (UniqueName: \"kubernetes.io/projected/a4f9a546-9052-4d89-8997-e826c9d8ed75-kube-api-access-v2rtg\") on node \"crc\" DevicePath \"\"" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.582843 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a546-9052-4d89-8997-e826c9d8ed75-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.608128 4735 scope.go:117] "RemoveContainer" containerID="f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.653860 4735 scope.go:117] "RemoveContainer" containerID="e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d" Nov 22 09:35:55 crc kubenswrapper[4735]: E1122 09:35:55.660129 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d\": container with ID starting with e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d not found: ID does not exist" containerID="e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.660399 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d"} err="failed to get container status \"e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d\": rpc error: code = NotFound desc = could not find container \"e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d\": container with ID starting with e83458f532310712d1b3192743d6518fc1b28b28217fef9b14c46f48d223f65d not found: ID does not exist" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.660519 4735 scope.go:117] "RemoveContainer" containerID="ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55" Nov 22 09:35:55 crc kubenswrapper[4735]: E1122 09:35:55.661132 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55\": container with ID starting with ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55 not found: ID does not exist" containerID="ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.661215 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55"} err="failed to get container status \"ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55\": rpc error: code = NotFound desc = could not find container \"ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55\": container with ID starting with ffb27591ebaefecc12d1ebfa645cfbdb0b2132e1627f58bb29be410e50e9fc55 not found: ID does not exist" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.661274 4735 scope.go:117] "RemoveContainer" containerID="f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3" Nov 22 09:35:55 crc kubenswrapper[4735]: E1122 09:35:55.661703 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3\": container with ID starting with f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3 not found: ID does not exist" containerID="f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.661812 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3"} err="failed to get container status \"f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3\": rpc error: code = NotFound desc = could not find container \"f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3\": container with ID starting with f678cbe2726705bc12594222f103a3d973ff486380fe535df80708827b8fc1d3 not found: ID does not exist" Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.893965 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9m58"] Nov 22 09:35:55 crc kubenswrapper[4735]: I1122 09:35:55.906612 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l9m58"] Nov 22 09:35:57 crc kubenswrapper[4735]: I1122 09:35:57.276835 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" path="/var/lib/kubelet/pods/a4f9a546-9052-4d89-8997-e826c9d8ed75/volumes" Nov 22 09:36:16 crc kubenswrapper[4735]: I1122 09:36:16.132190 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:36:16 crc kubenswrapper[4735]: I1122 09:36:16.132927 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:36:46 crc kubenswrapper[4735]: I1122 09:36:46.131836 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:36:46 crc kubenswrapper[4735]: I1122 09:36:46.132451 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.131609 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.132424 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.132523 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.134106 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.134237 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" gracePeriod=600 Nov 22 09:37:16 crc kubenswrapper[4735]: E1122 09:37:16.288595 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.571404 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" exitCode=0 Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.571453 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4"} Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.571522 4735 scope.go:117] "RemoveContainer" containerID="ab9ddf2de4a6cb93d17c04b33b3cf8f35dcb10a57394d20f90dceb88ba6ec85c" Nov 22 09:37:16 crc kubenswrapper[4735]: I1122 09:37:16.572156 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:37:16 crc kubenswrapper[4735]: E1122 09:37:16.572426 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:37:29 crc kubenswrapper[4735]: I1122 09:37:29.280334 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:37:29 crc kubenswrapper[4735]: E1122 09:37:29.281359 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:37:44 crc kubenswrapper[4735]: I1122 09:37:44.263705 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:37:44 crc kubenswrapper[4735]: E1122 09:37:44.264448 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:37:56 crc kubenswrapper[4735]: I1122 09:37:56.265136 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:37:56 crc kubenswrapper[4735]: E1122 09:37:56.265825 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:38:08 crc kubenswrapper[4735]: I1122 09:38:08.263238 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:38:08 crc kubenswrapper[4735]: E1122 09:38:08.263971 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:38:23 crc kubenswrapper[4735]: I1122 09:38:23.264530 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:38:23 crc kubenswrapper[4735]: E1122 09:38:23.265273 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:38:37 crc kubenswrapper[4735]: I1122 09:38:37.263986 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:38:37 crc kubenswrapper[4735]: E1122 09:38:37.265185 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:38:49 crc kubenswrapper[4735]: I1122 09:38:49.268107 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:38:49 crc kubenswrapper[4735]: E1122 09:38:49.268964 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:39:02 crc kubenswrapper[4735]: I1122 09:39:02.263199 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:39:02 crc kubenswrapper[4735]: E1122 09:39:02.263983 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:39:15 crc kubenswrapper[4735]: I1122 09:39:15.263618 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:39:15 crc kubenswrapper[4735]: E1122 09:39:15.264385 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:39:29 crc kubenswrapper[4735]: I1122 09:39:29.276042 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:39:29 crc kubenswrapper[4735]: E1122 09:39:29.277138 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:39:41 crc kubenswrapper[4735]: I1122 09:39:41.264337 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:39:41 crc kubenswrapper[4735]: E1122 09:39:41.265196 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:39:54 crc kubenswrapper[4735]: I1122 09:39:54.264363 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:39:54 crc kubenswrapper[4735]: E1122 09:39:54.265390 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:40:06 crc kubenswrapper[4735]: I1122 09:40:06.263191 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:40:06 crc kubenswrapper[4735]: E1122 09:40:06.264323 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:40:21 crc kubenswrapper[4735]: I1122 09:40:21.264577 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:40:21 crc kubenswrapper[4735]: E1122 09:40:21.265428 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:40:36 crc kubenswrapper[4735]: I1122 09:40:36.263689 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:40:36 crc kubenswrapper[4735]: E1122 09:40:36.264356 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:40:49 crc kubenswrapper[4735]: I1122 09:40:49.271413 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:40:49 crc kubenswrapper[4735]: E1122 09:40:49.272147 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.848221 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qf99q"] Nov 22 09:40:56 crc kubenswrapper[4735]: E1122 09:40:56.851829 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="extract-content" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.851864 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="extract-content" Nov 22 09:40:56 crc kubenswrapper[4735]: E1122 09:40:56.851883 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="extract-content" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.851891 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="extract-content" Nov 22 09:40:56 crc kubenswrapper[4735]: E1122 09:40:56.851924 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="registry-server" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.851931 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="registry-server" Nov 22 09:40:56 crc kubenswrapper[4735]: E1122 09:40:56.851953 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="registry-server" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.851958 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="registry-server" Nov 22 09:40:56 crc kubenswrapper[4735]: E1122 09:40:56.851972 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="extract-utilities" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.851979 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="extract-utilities" Nov 22 09:40:56 crc kubenswrapper[4735]: E1122 09:40:56.851992 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="extract-utilities" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.852000 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="extract-utilities" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.852215 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4f9a546-9052-4d89-8997-e826c9d8ed75" containerName="registry-server" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.852239 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ee3f179-5bf3-46a6-8980-0acfea0cd1cb" containerName="registry-server" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.854226 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.870602 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qf99q"] Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.979491 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctgsg\" (UniqueName: \"kubernetes.io/projected/ff6e281d-b2b1-4d4b-a148-5986e321cdde-kube-api-access-ctgsg\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.980563 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-catalog-content\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:56 crc kubenswrapper[4735]: I1122 09:40:56.980762 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-utilities\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:57 crc kubenswrapper[4735]: I1122 09:40:57.083952 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-catalog-content\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:57 crc kubenswrapper[4735]: I1122 09:40:57.084080 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-utilities\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:57 crc kubenswrapper[4735]: I1122 09:40:57.084333 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctgsg\" (UniqueName: \"kubernetes.io/projected/ff6e281d-b2b1-4d4b-a148-5986e321cdde-kube-api-access-ctgsg\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:57 crc kubenswrapper[4735]: I1122 09:40:57.084952 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-utilities\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:57 crc kubenswrapper[4735]: I1122 09:40:57.084987 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-catalog-content\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:57 crc kubenswrapper[4735]: I1122 09:40:57.113342 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctgsg\" (UniqueName: \"kubernetes.io/projected/ff6e281d-b2b1-4d4b-a148-5986e321cdde-kube-api-access-ctgsg\") pod \"redhat-marketplace-qf99q\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:57 crc kubenswrapper[4735]: I1122 09:40:57.185302 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:40:57 crc kubenswrapper[4735]: I1122 09:40:57.685925 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qf99q"] Nov 22 09:40:58 crc kubenswrapper[4735]: I1122 09:40:58.183495 4735 generic.go:334] "Generic (PLEG): container finished" podID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerID="f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17" exitCode=0 Nov 22 09:40:58 crc kubenswrapper[4735]: I1122 09:40:58.183715 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qf99q" event={"ID":"ff6e281d-b2b1-4d4b-a148-5986e321cdde","Type":"ContainerDied","Data":"f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17"} Nov 22 09:40:58 crc kubenswrapper[4735]: I1122 09:40:58.183833 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qf99q" event={"ID":"ff6e281d-b2b1-4d4b-a148-5986e321cdde","Type":"ContainerStarted","Data":"14e546074d94186c6e4ae5032084607a5a460435e80cbfda5cf3e7d98f411bd1"} Nov 22 09:40:58 crc kubenswrapper[4735]: I1122 09:40:58.185564 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:40:59 crc kubenswrapper[4735]: I1122 09:40:59.196163 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qf99q" event={"ID":"ff6e281d-b2b1-4d4b-a148-5986e321cdde","Type":"ContainerStarted","Data":"d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680"} Nov 22 09:41:00 crc kubenswrapper[4735]: I1122 09:41:00.210121 4735 generic.go:334] "Generic (PLEG): container finished" podID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerID="d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680" exitCode=0 Nov 22 09:41:00 crc kubenswrapper[4735]: I1122 09:41:00.210225 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qf99q" event={"ID":"ff6e281d-b2b1-4d4b-a148-5986e321cdde","Type":"ContainerDied","Data":"d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680"} Nov 22 09:41:01 crc kubenswrapper[4735]: I1122 09:41:01.226346 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qf99q" event={"ID":"ff6e281d-b2b1-4d4b-a148-5986e321cdde","Type":"ContainerStarted","Data":"414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe"} Nov 22 09:41:01 crc kubenswrapper[4735]: I1122 09:41:01.253258 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qf99q" podStartSLOduration=2.795848018 podStartE2EDuration="5.253237163s" podCreationTimestamp="2025-11-22 09:40:56 +0000 UTC" firstStartedPulling="2025-11-22 09:40:58.18526701 +0000 UTC m=+5879.789605615" lastFinishedPulling="2025-11-22 09:41:00.642656145 +0000 UTC m=+5882.246994760" observedRunningTime="2025-11-22 09:41:01.250342324 +0000 UTC m=+5882.854680929" watchObservedRunningTime="2025-11-22 09:41:01.253237163 +0000 UTC m=+5882.857575768" Nov 22 09:41:03 crc kubenswrapper[4735]: I1122 09:41:03.263555 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:41:03 crc kubenswrapper[4735]: E1122 09:41:03.264223 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:41:07 crc kubenswrapper[4735]: I1122 09:41:07.186428 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:41:07 crc kubenswrapper[4735]: I1122 09:41:07.186927 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:41:07 crc kubenswrapper[4735]: I1122 09:41:07.288558 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:41:07 crc kubenswrapper[4735]: I1122 09:41:07.361063 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:41:07 crc kubenswrapper[4735]: I1122 09:41:07.533970 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qf99q"] Nov 22 09:41:09 crc kubenswrapper[4735]: I1122 09:41:09.325594 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qf99q" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerName="registry-server" containerID="cri-o://414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe" gracePeriod=2 Nov 22 09:41:09 crc kubenswrapper[4735]: I1122 09:41:09.915768 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.106822 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-catalog-content\") pod \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.106900 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-utilities\") pod \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.106976 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctgsg\" (UniqueName: \"kubernetes.io/projected/ff6e281d-b2b1-4d4b-a148-5986e321cdde-kube-api-access-ctgsg\") pod \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\" (UID: \"ff6e281d-b2b1-4d4b-a148-5986e321cdde\") " Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.108366 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-utilities" (OuterVolumeSpecName: "utilities") pod "ff6e281d-b2b1-4d4b-a148-5986e321cdde" (UID: "ff6e281d-b2b1-4d4b-a148-5986e321cdde"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.120357 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff6e281d-b2b1-4d4b-a148-5986e321cdde-kube-api-access-ctgsg" (OuterVolumeSpecName: "kube-api-access-ctgsg") pod "ff6e281d-b2b1-4d4b-a148-5986e321cdde" (UID: "ff6e281d-b2b1-4d4b-a148-5986e321cdde"). InnerVolumeSpecName "kube-api-access-ctgsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.124544 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff6e281d-b2b1-4d4b-a148-5986e321cdde" (UID: "ff6e281d-b2b1-4d4b-a148-5986e321cdde"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.212181 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.212217 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6e281d-b2b1-4d4b-a148-5986e321cdde-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.212395 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctgsg\" (UniqueName: \"kubernetes.io/projected/ff6e281d-b2b1-4d4b-a148-5986e321cdde-kube-api-access-ctgsg\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.338069 4735 generic.go:334] "Generic (PLEG): container finished" podID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerID="414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe" exitCode=0 Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.338150 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qf99q" event={"ID":"ff6e281d-b2b1-4d4b-a148-5986e321cdde","Type":"ContainerDied","Data":"414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe"} Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.338173 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qf99q" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.339511 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qf99q" event={"ID":"ff6e281d-b2b1-4d4b-a148-5986e321cdde","Type":"ContainerDied","Data":"14e546074d94186c6e4ae5032084607a5a460435e80cbfda5cf3e7d98f411bd1"} Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.339622 4735 scope.go:117] "RemoveContainer" containerID="414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.366119 4735 scope.go:117] "RemoveContainer" containerID="d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.407505 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qf99q"] Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.433654 4735 scope.go:117] "RemoveContainer" containerID="f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.478646 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qf99q"] Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.501315 4735 scope.go:117] "RemoveContainer" containerID="414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe" Nov 22 09:41:10 crc kubenswrapper[4735]: E1122 09:41:10.502282 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe\": container with ID starting with 414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe not found: ID does not exist" containerID="414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.502324 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe"} err="failed to get container status \"414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe\": rpc error: code = NotFound desc = could not find container \"414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe\": container with ID starting with 414c4f92a6bde6907234de8b36fa4e8587285348ee6b25ecf65cc429f6a5f6fe not found: ID does not exist" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.502351 4735 scope.go:117] "RemoveContainer" containerID="d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680" Nov 22 09:41:10 crc kubenswrapper[4735]: E1122 09:41:10.502761 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680\": container with ID starting with d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680 not found: ID does not exist" containerID="d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.502788 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680"} err="failed to get container status \"d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680\": rpc error: code = NotFound desc = could not find container \"d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680\": container with ID starting with d7f1211ef69dcab0df9cc53dafdfef287fc10f736dc6f3a21376dfa0eacda680 not found: ID does not exist" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.502807 4735 scope.go:117] "RemoveContainer" containerID="f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17" Nov 22 09:41:10 crc kubenswrapper[4735]: E1122 09:41:10.503018 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17\": container with ID starting with f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17 not found: ID does not exist" containerID="f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17" Nov 22 09:41:10 crc kubenswrapper[4735]: I1122 09:41:10.503042 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17"} err="failed to get container status \"f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17\": rpc error: code = NotFound desc = could not find container \"f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17\": container with ID starting with f85eebe6791d266029dac80f614ceda841554b2b264192afa41b8b9e8f50fa17 not found: ID does not exist" Nov 22 09:41:11 crc kubenswrapper[4735]: I1122 09:41:11.280050 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" path="/var/lib/kubelet/pods/ff6e281d-b2b1-4d4b-a148-5986e321cdde/volumes" Nov 22 09:41:16 crc kubenswrapper[4735]: I1122 09:41:16.264396 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:41:16 crc kubenswrapper[4735]: E1122 09:41:16.265432 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:41:23 crc kubenswrapper[4735]: I1122 09:41:23.492253 4735 generic.go:334] "Generic (PLEG): container finished" podID="9d89aa19-7bdb-4db2-ad64-438da4087c36" containerID="a4f2b24b4ad681dbaf10f89723ebe5e7d6b16ef9244464d9351710b372ccee1a" exitCode=0 Nov 22 09:41:23 crc kubenswrapper[4735]: I1122 09:41:23.492359 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9d89aa19-7bdb-4db2-ad64-438da4087c36","Type":"ContainerDied","Data":"a4f2b24b4ad681dbaf10f89723ebe5e7d6b16ef9244464d9351710b372ccee1a"} Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.898434 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910274 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config-secret\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910437 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v72xk\" (UniqueName: \"kubernetes.io/projected/9d89aa19-7bdb-4db2-ad64-438da4087c36-kube-api-access-v72xk\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910599 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ssh-key\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910641 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ca-certs\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910700 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910767 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-config-data\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910804 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910828 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-workdir\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.910914 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-temporary\") pod \"9d89aa19-7bdb-4db2-ad64-438da4087c36\" (UID: \"9d89aa19-7bdb-4db2-ad64-438da4087c36\") " Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.911903 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-config-data" (OuterVolumeSpecName: "config-data") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.912316 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.912493 4735 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.912512 4735 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.926146 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d89aa19-7bdb-4db2-ad64-438da4087c36-kube-api-access-v72xk" (OuterVolumeSpecName: "kube-api-access-v72xk") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "kube-api-access-v72xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.927585 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.929694 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.960075 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.967962 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:41:24 crc kubenswrapper[4735]: I1122 09:41:24.997264 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.007229 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9d89aa19-7bdb-4db2-ad64-438da4087c36" (UID: "9d89aa19-7bdb-4db2-ad64-438da4087c36"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.016518 4735 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.016553 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.016567 4735 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9d89aa19-7bdb-4db2-ad64-438da4087c36-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.016578 4735 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.016589 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v72xk\" (UniqueName: \"kubernetes.io/projected/9d89aa19-7bdb-4db2-ad64-438da4087c36-kube-api-access-v72xk\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.016598 4735 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.016606 4735 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9d89aa19-7bdb-4db2-ad64-438da4087c36-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.044257 4735 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.119973 4735 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.522224 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9d89aa19-7bdb-4db2-ad64-438da4087c36","Type":"ContainerDied","Data":"1a43c20d6071e440ed3eb2a2dc72346e6ca3feb2cfbc7219f2d9ceccebbf61be"} Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.522570 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a43c20d6071e440ed3eb2a2dc72346e6ca3feb2cfbc7219f2d9ceccebbf61be" Nov 22 09:41:25 crc kubenswrapper[4735]: I1122 09:41:25.522275 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 22 09:41:29 crc kubenswrapper[4735]: I1122 09:41:29.271253 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:41:29 crc kubenswrapper[4735]: E1122 09:41:29.272221 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.907940 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 09:41:36 crc kubenswrapper[4735]: E1122 09:41:36.908957 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerName="extract-utilities" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.908973 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerName="extract-utilities" Nov 22 09:41:36 crc kubenswrapper[4735]: E1122 09:41:36.908991 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerName="registry-server" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.908997 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerName="registry-server" Nov 22 09:41:36 crc kubenswrapper[4735]: E1122 09:41:36.909015 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d89aa19-7bdb-4db2-ad64-438da4087c36" containerName="tempest-tests-tempest-tests-runner" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.909023 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d89aa19-7bdb-4db2-ad64-438da4087c36" containerName="tempest-tests-tempest-tests-runner" Nov 22 09:41:36 crc kubenswrapper[4735]: E1122 09:41:36.909045 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerName="extract-content" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.909051 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerName="extract-content" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.909264 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff6e281d-b2b1-4d4b-a148-5986e321cdde" containerName="registry-server" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.909273 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d89aa19-7bdb-4db2-ad64-438da4087c36" containerName="tempest-tests-tempest-tests-runner" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.910101 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.914570 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-2lmrg" Nov 22 09:41:36 crc kubenswrapper[4735]: I1122 09:41:36.918877 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 09:41:37 crc kubenswrapper[4735]: I1122 09:41:37.112136 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff52r\" (UniqueName: \"kubernetes.io/projected/c68fad70-fe37-44d9-9fce-83f3e4c5d55a-kube-api-access-ff52r\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c68fad70-fe37-44d9-9fce-83f3e4c5d55a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:37 crc kubenswrapper[4735]: I1122 09:41:37.112785 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c68fad70-fe37-44d9-9fce-83f3e4c5d55a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:37 crc kubenswrapper[4735]: I1122 09:41:37.214623 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c68fad70-fe37-44d9-9fce-83f3e4c5d55a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:37 crc kubenswrapper[4735]: I1122 09:41:37.214946 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff52r\" (UniqueName: \"kubernetes.io/projected/c68fad70-fe37-44d9-9fce-83f3e4c5d55a-kube-api-access-ff52r\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c68fad70-fe37-44d9-9fce-83f3e4c5d55a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:37 crc kubenswrapper[4735]: I1122 09:41:37.216339 4735 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c68fad70-fe37-44d9-9fce-83f3e4c5d55a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:37 crc kubenswrapper[4735]: I1122 09:41:37.235874 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff52r\" (UniqueName: \"kubernetes.io/projected/c68fad70-fe37-44d9-9fce-83f3e4c5d55a-kube-api-access-ff52r\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c68fad70-fe37-44d9-9fce-83f3e4c5d55a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:37 crc kubenswrapper[4735]: I1122 09:41:37.270340 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c68fad70-fe37-44d9-9fce-83f3e4c5d55a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:37 crc kubenswrapper[4735]: I1122 09:41:37.531837 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 09:41:38 crc kubenswrapper[4735]: I1122 09:41:38.031035 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 09:41:38 crc kubenswrapper[4735]: I1122 09:41:38.693337 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c68fad70-fe37-44d9-9fce-83f3e4c5d55a","Type":"ContainerStarted","Data":"3e5ef52f253d7cdc80dc9c735d0537f3c4184864bc4451e7189b1c5316c0c0d7"} Nov 22 09:41:40 crc kubenswrapper[4735]: I1122 09:41:40.721786 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c68fad70-fe37-44d9-9fce-83f3e4c5d55a","Type":"ContainerStarted","Data":"d2e2c18b57f6b18b3392d9f697d1e4d9918b8f843c6016e1564c981e143c89b0"} Nov 22 09:41:40 crc kubenswrapper[4735]: I1122 09:41:40.742520 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.158678122 podStartE2EDuration="4.74249259s" podCreationTimestamp="2025-11-22 09:41:36 +0000 UTC" firstStartedPulling="2025-11-22 09:41:38.042119818 +0000 UTC m=+5919.646458423" lastFinishedPulling="2025-11-22 09:41:39.625934286 +0000 UTC m=+5921.230272891" observedRunningTime="2025-11-22 09:41:40.736822344 +0000 UTC m=+5922.341160949" watchObservedRunningTime="2025-11-22 09:41:40.74249259 +0000 UTC m=+5922.346831225" Nov 22 09:41:43 crc kubenswrapper[4735]: I1122 09:41:43.263672 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:41:43 crc kubenswrapper[4735]: E1122 09:41:43.264772 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:41:54 crc kubenswrapper[4735]: I1122 09:41:54.264933 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:41:54 crc kubenswrapper[4735]: E1122 09:41:54.266077 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:42:08 crc kubenswrapper[4735]: I1122 09:42:08.264260 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:42:08 crc kubenswrapper[4735]: E1122 09:42:08.265028 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.070375 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ldm2z/must-gather-npq4c"] Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.072735 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.074450 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ldm2z"/"default-dockercfg-gd6cd" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.074709 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ldm2z"/"kube-root-ca.crt" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.074741 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ldm2z"/"openshift-service-ca.crt" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.086027 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ldm2z/must-gather-npq4c"] Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.137027 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbc2w\" (UniqueName: \"kubernetes.io/projected/c31c2609-9e07-4d74-b13a-8af5097f662c-kube-api-access-fbc2w\") pod \"must-gather-npq4c\" (UID: \"c31c2609-9e07-4d74-b13a-8af5097f662c\") " pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.137146 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c31c2609-9e07-4d74-b13a-8af5097f662c-must-gather-output\") pod \"must-gather-npq4c\" (UID: \"c31c2609-9e07-4d74-b13a-8af5097f662c\") " pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.239109 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbc2w\" (UniqueName: \"kubernetes.io/projected/c31c2609-9e07-4d74-b13a-8af5097f662c-kube-api-access-fbc2w\") pod \"must-gather-npq4c\" (UID: \"c31c2609-9e07-4d74-b13a-8af5097f662c\") " pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.239189 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c31c2609-9e07-4d74-b13a-8af5097f662c-must-gather-output\") pod \"must-gather-npq4c\" (UID: \"c31c2609-9e07-4d74-b13a-8af5097f662c\") " pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.239651 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c31c2609-9e07-4d74-b13a-8af5097f662c-must-gather-output\") pod \"must-gather-npq4c\" (UID: \"c31c2609-9e07-4d74-b13a-8af5097f662c\") " pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.256325 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbc2w\" (UniqueName: \"kubernetes.io/projected/c31c2609-9e07-4d74-b13a-8af5097f662c-kube-api-access-fbc2w\") pod \"must-gather-npq4c\" (UID: \"c31c2609-9e07-4d74-b13a-8af5097f662c\") " pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.391707 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:17 crc kubenswrapper[4735]: I1122 09:42:17.872086 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ldm2z/must-gather-npq4c"] Nov 22 09:42:18 crc kubenswrapper[4735]: I1122 09:42:18.181338 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldm2z/must-gather-npq4c" event={"ID":"c31c2609-9e07-4d74-b13a-8af5097f662c","Type":"ContainerStarted","Data":"c847e9d60492708074e9f1bfbcaa3d98c99160b3874e56abfd17eca65cfdad50"} Nov 22 09:42:20 crc kubenswrapper[4735]: I1122 09:42:20.264115 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:42:21 crc kubenswrapper[4735]: I1122 09:42:21.215509 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"76f202ccd08e999858681fe56c8b8a5850e9e3360e4ffa862758f705684ac588"} Nov 22 09:42:31 crc kubenswrapper[4735]: I1122 09:42:31.179744 4735 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.369071092s: [/var/lib/containers/storage/overlay/02dc9a766aa522c4a1efcf91c0e5b285680d0978f265a78e7d8e712e555cc83c/diff /var/log/pods/openshift-network-console_networking-console-plugin-85b44fc459-gdk6g_5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/networking-console-plugin/0.log]; will not log again for this container unless duration exceeds 2s Nov 22 09:42:34 crc kubenswrapper[4735]: E1122 09:42:34.738706 4735 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-must-gather:latest" Nov 22 09:42:34 crc kubenswrapper[4735]: E1122 09:42:34.763444 4735 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 09:42:34 crc kubenswrapper[4735]: container &Container{Name:gather,Image:quay.io/openstack-k8s-operators/openstack-must-gather:latest,Command:[/bin/bash -c Nov 22 09:42:34 crc kubenswrapper[4735]: echo "[disk usage checker] Started" Nov 22 09:42:34 crc kubenswrapper[4735]: target_dir="/must-gather" Nov 22 09:42:34 crc kubenswrapper[4735]: usage_percentage_limit="30" Nov 22 09:42:34 crc kubenswrapper[4735]: while true; do Nov 22 09:42:34 crc kubenswrapper[4735]: disk_usage=$(du -s "$target_dir" | awk '{print $1}') Nov 22 09:42:34 crc kubenswrapper[4735]: disk_space=$(df -P "$target_dir" | awk 'NR==2 {print $2}') Nov 22 09:42:34 crc kubenswrapper[4735]: usage_percentage=$(( (disk_usage * 100) / disk_space )) Nov 22 09:42:34 crc kubenswrapper[4735]: echo "[disk usage checker] Volume usage percentage: current = ${usage_percentage} ; allowed = ${usage_percentage_limit}" Nov 22 09:42:34 crc kubenswrapper[4735]: if [ "$usage_percentage" -gt "$usage_percentage_limit" ]; then Nov 22 09:42:34 crc kubenswrapper[4735]: echo "[disk usage checker] Disk usage exceeds the volume percentage of ${usage_percentage_limit} for mounted directory, terminating..." Nov 22 09:42:34 crc kubenswrapper[4735]: pkill --signal SIGKILL gather Nov 22 09:42:34 crc kubenswrapper[4735]: exit 1 Nov 22 09:42:34 crc kubenswrapper[4735]: fi Nov 22 09:42:34 crc kubenswrapper[4735]: sleep 5 Nov 22 09:42:34 crc kubenswrapper[4735]: done & ADDITIONAL_NAMESPACES=kuttl,openshift-storage,openshift-marketplace,openshift-operators,sushy-emulator,tobiko OPENSTACK_DATABASES=ALL SOS_EDPM=all SOS_DECOMPRESS=0 gather; sync && echo 'Caches written to disk'],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:must-gather-output,ReadOnly:false,MountPath:/must-gather,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fbc2w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod must-gather-npq4c_openshift-must-gather-ldm2z(c31c2609-9e07-4d74-b13a-8af5097f662c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 22 09:42:34 crc kubenswrapper[4735]: > logger="UnhandledError" Nov 22 09:42:34 crc kubenswrapper[4735]: E1122 09:42:34.766923 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"gather\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"copy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\"]" pod="openshift-must-gather-ldm2z/must-gather-npq4c" podUID="c31c2609-9e07-4d74-b13a-8af5097f662c" Nov 22 09:42:35 crc kubenswrapper[4735]: E1122 09:42:35.431777 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"gather\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\", failed to \"StartContainer\" for \"copy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\"]" pod="openshift-must-gather-ldm2z/must-gather-npq4c" podUID="c31c2609-9e07-4d74-b13a-8af5097f662c" Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.206937 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ldm2z/must-gather-npq4c"] Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.234129 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ldm2z/must-gather-npq4c"] Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.630676 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.702189 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbc2w\" (UniqueName: \"kubernetes.io/projected/c31c2609-9e07-4d74-b13a-8af5097f662c-kube-api-access-fbc2w\") pod \"c31c2609-9e07-4d74-b13a-8af5097f662c\" (UID: \"c31c2609-9e07-4d74-b13a-8af5097f662c\") " Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.702425 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c31c2609-9e07-4d74-b13a-8af5097f662c-must-gather-output\") pod \"c31c2609-9e07-4d74-b13a-8af5097f662c\" (UID: \"c31c2609-9e07-4d74-b13a-8af5097f662c\") " Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.702788 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c31c2609-9e07-4d74-b13a-8af5097f662c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c31c2609-9e07-4d74-b13a-8af5097f662c" (UID: "c31c2609-9e07-4d74-b13a-8af5097f662c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.703144 4735 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c31c2609-9e07-4d74-b13a-8af5097f662c-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.711088 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31c2609-9e07-4d74-b13a-8af5097f662c-kube-api-access-fbc2w" (OuterVolumeSpecName: "kube-api-access-fbc2w") pod "c31c2609-9e07-4d74-b13a-8af5097f662c" (UID: "c31c2609-9e07-4d74-b13a-8af5097f662c"). InnerVolumeSpecName "kube-api-access-fbc2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:42:42 crc kubenswrapper[4735]: I1122 09:42:42.805069 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbc2w\" (UniqueName: \"kubernetes.io/projected/c31c2609-9e07-4d74-b13a-8af5097f662c-kube-api-access-fbc2w\") on node \"crc\" DevicePath \"\"" Nov 22 09:42:43 crc kubenswrapper[4735]: I1122 09:42:43.276177 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c31c2609-9e07-4d74-b13a-8af5097f662c" path="/var/lib/kubelet/pods/c31c2609-9e07-4d74-b13a-8af5097f662c/volumes" Nov 22 09:42:43 crc kubenswrapper[4735]: I1122 09:42:43.528746 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldm2z/must-gather-npq4c" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.455621 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xgm58/must-gather-vrnsx"] Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.458498 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.465222 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xgm58"/"kube-root-ca.crt" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.475488 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xgm58"/"openshift-service-ca.crt" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.482010 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xgm58/must-gather-vrnsx"] Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.620498 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-must-gather-output\") pod \"must-gather-vrnsx\" (UID: \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\") " pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.620651 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ddbf\" (UniqueName: \"kubernetes.io/projected/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-kube-api-access-2ddbf\") pod \"must-gather-vrnsx\" (UID: \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\") " pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.722685 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ddbf\" (UniqueName: \"kubernetes.io/projected/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-kube-api-access-2ddbf\") pod \"must-gather-vrnsx\" (UID: \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\") " pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.722895 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-must-gather-output\") pod \"must-gather-vrnsx\" (UID: \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\") " pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.723364 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-must-gather-output\") pod \"must-gather-vrnsx\" (UID: \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\") " pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.744349 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ddbf\" (UniqueName: \"kubernetes.io/projected/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-kube-api-access-2ddbf\") pod \"must-gather-vrnsx\" (UID: \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\") " pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:43:31 crc kubenswrapper[4735]: I1122 09:43:31.789378 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:43:32 crc kubenswrapper[4735]: I1122 09:43:32.491253 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xgm58/must-gather-vrnsx"] Nov 22 09:43:33 crc kubenswrapper[4735]: I1122 09:43:33.198752 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/must-gather-vrnsx" event={"ID":"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0","Type":"ContainerStarted","Data":"d249382f0fdc529bfe281e3ac3c416cd39bcf7af98fe5730f3517d3d4b71457e"} Nov 22 09:43:34 crc kubenswrapper[4735]: I1122 09:43:34.217533 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/must-gather-vrnsx" event={"ID":"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0","Type":"ContainerStarted","Data":"124998c2c39ad7430659eecf4bcc3ee5ea6a0b4b4969547a5238877e54d7441c"} Nov 22 09:43:34 crc kubenswrapper[4735]: I1122 09:43:34.220244 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/must-gather-vrnsx" event={"ID":"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0","Type":"ContainerStarted","Data":"e52d195a536f7c64458497cd7b12a33af17abdc43c79bccdc97c518b3c2d9c5c"} Nov 22 09:43:34 crc kubenswrapper[4735]: I1122 09:43:34.246596 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xgm58/must-gather-vrnsx" podStartSLOduration=2.795626022 podStartE2EDuration="3.246573769s" podCreationTimestamp="2025-11-22 09:43:31 +0000 UTC" firstStartedPulling="2025-11-22 09:43:32.498502966 +0000 UTC m=+6034.102841611" lastFinishedPulling="2025-11-22 09:43:32.949450743 +0000 UTC m=+6034.553789358" observedRunningTime="2025-11-22 09:43:34.23676632 +0000 UTC m=+6035.841104945" watchObservedRunningTime="2025-11-22 09:43:34.246573769 +0000 UTC m=+6035.850912394" Nov 22 09:43:38 crc kubenswrapper[4735]: I1122 09:43:38.965615 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xgm58/crc-debug-2cn79"] Nov 22 09:43:38 crc kubenswrapper[4735]: I1122 09:43:38.968007 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:43:38 crc kubenswrapper[4735]: I1122 09:43:38.978863 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xgm58"/"default-dockercfg-tfbj6" Nov 22 09:43:39 crc kubenswrapper[4735]: I1122 09:43:39.112935 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkqxl\" (UniqueName: \"kubernetes.io/projected/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-kube-api-access-qkqxl\") pod \"crc-debug-2cn79\" (UID: \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\") " pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:43:39 crc kubenswrapper[4735]: I1122 09:43:39.113847 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-host\") pod \"crc-debug-2cn79\" (UID: \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\") " pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:43:39 crc kubenswrapper[4735]: I1122 09:43:39.216098 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-host\") pod \"crc-debug-2cn79\" (UID: \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\") " pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:43:39 crc kubenswrapper[4735]: I1122 09:43:39.216180 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkqxl\" (UniqueName: \"kubernetes.io/projected/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-kube-api-access-qkqxl\") pod \"crc-debug-2cn79\" (UID: \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\") " pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:43:39 crc kubenswrapper[4735]: I1122 09:43:39.217185 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-host\") pod \"crc-debug-2cn79\" (UID: \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\") " pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:43:39 crc kubenswrapper[4735]: I1122 09:43:39.238535 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkqxl\" (UniqueName: \"kubernetes.io/projected/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-kube-api-access-qkqxl\") pod \"crc-debug-2cn79\" (UID: \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\") " pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:43:39 crc kubenswrapper[4735]: I1122 09:43:39.311329 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:43:40 crc kubenswrapper[4735]: I1122 09:43:40.284871 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/crc-debug-2cn79" event={"ID":"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac","Type":"ContainerStarted","Data":"f53c14ea3521f4b1fdc3dfd018ae8cfd03f77e37f839f70720538c26188a49f7"} Nov 22 09:43:51 crc kubenswrapper[4735]: I1122 09:43:51.439187 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/crc-debug-2cn79" event={"ID":"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac","Type":"ContainerStarted","Data":"edab236d01911533c73a2289a0328c4b650ab4bbcf8be85ac0e43bf6be7c1bfb"} Nov 22 09:43:51 crc kubenswrapper[4735]: I1122 09:43:51.464138 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xgm58/crc-debug-2cn79" podStartSLOduration=1.963036345 podStartE2EDuration="13.464118303s" podCreationTimestamp="2025-11-22 09:43:38 +0000 UTC" firstStartedPulling="2025-11-22 09:43:39.389121707 +0000 UTC m=+6040.993460312" lastFinishedPulling="2025-11-22 09:43:50.890203665 +0000 UTC m=+6052.494542270" observedRunningTime="2025-11-22 09:43:51.462731195 +0000 UTC m=+6053.067069800" watchObservedRunningTime="2025-11-22 09:43:51.464118303 +0000 UTC m=+6053.068456908" Nov 22 09:44:44 crc kubenswrapper[4735]: I1122 09:44:44.112382 4735 generic.go:334] "Generic (PLEG): container finished" podID="5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac" containerID="edab236d01911533c73a2289a0328c4b650ab4bbcf8be85ac0e43bf6be7c1bfb" exitCode=0 Nov 22 09:44:44 crc kubenswrapper[4735]: I1122 09:44:44.112430 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/crc-debug-2cn79" event={"ID":"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac","Type":"ContainerDied","Data":"edab236d01911533c73a2289a0328c4b650ab4bbcf8be85ac0e43bf6be7c1bfb"} Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.289509 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.340492 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xgm58/crc-debug-2cn79"] Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.353434 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xgm58/crc-debug-2cn79"] Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.433765 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkqxl\" (UniqueName: \"kubernetes.io/projected/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-kube-api-access-qkqxl\") pod \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\" (UID: \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\") " Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.433853 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-host\") pod \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\" (UID: \"5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac\") " Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.434102 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-host" (OuterVolumeSpecName: "host") pod "5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac" (UID: "5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.434886 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-host\") on node \"crc\" DevicePath \"\"" Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.444073 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-kube-api-access-qkqxl" (OuterVolumeSpecName: "kube-api-access-qkqxl") pod "5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac" (UID: "5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac"). InnerVolumeSpecName "kube-api-access-qkqxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:44:45 crc kubenswrapper[4735]: I1122 09:44:45.537014 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkqxl\" (UniqueName: \"kubernetes.io/projected/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac-kube-api-access-qkqxl\") on node \"crc\" DevicePath \"\"" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.131764 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.131846 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.142618 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f53c14ea3521f4b1fdc3dfd018ae8cfd03f77e37f839f70720538c26188a49f7" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.142688 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-2cn79" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.537096 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xgm58/crc-debug-hj44k"] Nov 22 09:44:46 crc kubenswrapper[4735]: E1122 09:44:46.537988 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac" containerName="container-00" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.538007 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac" containerName="container-00" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.538378 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac" containerName="container-00" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.539374 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.541834 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xgm58"/"default-dockercfg-tfbj6" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.663946 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqlm8\" (UniqueName: \"kubernetes.io/projected/70281419-fcb7-434d-bc93-eb3cc918f1da-kube-api-access-fqlm8\") pod \"crc-debug-hj44k\" (UID: \"70281419-fcb7-434d-bc93-eb3cc918f1da\") " pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.664385 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70281419-fcb7-434d-bc93-eb3cc918f1da-host\") pod \"crc-debug-hj44k\" (UID: \"70281419-fcb7-434d-bc93-eb3cc918f1da\") " pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.766860 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqlm8\" (UniqueName: \"kubernetes.io/projected/70281419-fcb7-434d-bc93-eb3cc918f1da-kube-api-access-fqlm8\") pod \"crc-debug-hj44k\" (UID: \"70281419-fcb7-434d-bc93-eb3cc918f1da\") " pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.767064 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70281419-fcb7-434d-bc93-eb3cc918f1da-host\") pod \"crc-debug-hj44k\" (UID: \"70281419-fcb7-434d-bc93-eb3cc918f1da\") " pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.767202 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70281419-fcb7-434d-bc93-eb3cc918f1da-host\") pod \"crc-debug-hj44k\" (UID: \"70281419-fcb7-434d-bc93-eb3cc918f1da\") " pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.791701 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqlm8\" (UniqueName: \"kubernetes.io/projected/70281419-fcb7-434d-bc93-eb3cc918f1da-kube-api-access-fqlm8\") pod \"crc-debug-hj44k\" (UID: \"70281419-fcb7-434d-bc93-eb3cc918f1da\") " pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:46 crc kubenswrapper[4735]: I1122 09:44:46.864197 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:47 crc kubenswrapper[4735]: I1122 09:44:47.153472 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/crc-debug-hj44k" event={"ID":"70281419-fcb7-434d-bc93-eb3cc918f1da","Type":"ContainerStarted","Data":"c6d34f31784d360c04f48988d599d33eda626ec5cf1280ed61ceb934ff34ab31"} Nov 22 09:44:47 crc kubenswrapper[4735]: I1122 09:44:47.282227 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac" path="/var/lib/kubelet/pods/5c15fbb2-8dd0-4e5a-b332-f43f2afc5dac/volumes" Nov 22 09:44:48 crc kubenswrapper[4735]: I1122 09:44:48.169385 4735 generic.go:334] "Generic (PLEG): container finished" podID="70281419-fcb7-434d-bc93-eb3cc918f1da" containerID="4ca5b0cca0066f58a42fe9588104a7fb2f6d9a4addaf5fca99ef7fa65936fbf1" exitCode=0 Nov 22 09:44:48 crc kubenswrapper[4735]: I1122 09:44:48.169449 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/crc-debug-hj44k" event={"ID":"70281419-fcb7-434d-bc93-eb3cc918f1da","Type":"ContainerDied","Data":"4ca5b0cca0066f58a42fe9588104a7fb2f6d9a4addaf5fca99ef7fa65936fbf1"} Nov 22 09:44:49 crc kubenswrapper[4735]: I1122 09:44:49.294488 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:49 crc kubenswrapper[4735]: I1122 09:44:49.433672 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqlm8\" (UniqueName: \"kubernetes.io/projected/70281419-fcb7-434d-bc93-eb3cc918f1da-kube-api-access-fqlm8\") pod \"70281419-fcb7-434d-bc93-eb3cc918f1da\" (UID: \"70281419-fcb7-434d-bc93-eb3cc918f1da\") " Nov 22 09:44:49 crc kubenswrapper[4735]: I1122 09:44:49.433926 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70281419-fcb7-434d-bc93-eb3cc918f1da-host\") pod \"70281419-fcb7-434d-bc93-eb3cc918f1da\" (UID: \"70281419-fcb7-434d-bc93-eb3cc918f1da\") " Nov 22 09:44:49 crc kubenswrapper[4735]: I1122 09:44:49.434451 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70281419-fcb7-434d-bc93-eb3cc918f1da-host" (OuterVolumeSpecName: "host") pod "70281419-fcb7-434d-bc93-eb3cc918f1da" (UID: "70281419-fcb7-434d-bc93-eb3cc918f1da"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 09:44:49 crc kubenswrapper[4735]: I1122 09:44:49.434740 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70281419-fcb7-434d-bc93-eb3cc918f1da-host\") on node \"crc\" DevicePath \"\"" Nov 22 09:44:49 crc kubenswrapper[4735]: I1122 09:44:49.440510 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70281419-fcb7-434d-bc93-eb3cc918f1da-kube-api-access-fqlm8" (OuterVolumeSpecName: "kube-api-access-fqlm8") pod "70281419-fcb7-434d-bc93-eb3cc918f1da" (UID: "70281419-fcb7-434d-bc93-eb3cc918f1da"). InnerVolumeSpecName "kube-api-access-fqlm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:44:49 crc kubenswrapper[4735]: I1122 09:44:49.537013 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqlm8\" (UniqueName: \"kubernetes.io/projected/70281419-fcb7-434d-bc93-eb3cc918f1da-kube-api-access-fqlm8\") on node \"crc\" DevicePath \"\"" Nov 22 09:44:50 crc kubenswrapper[4735]: I1122 09:44:50.193821 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/crc-debug-hj44k" event={"ID":"70281419-fcb7-434d-bc93-eb3cc918f1da","Type":"ContainerDied","Data":"c6d34f31784d360c04f48988d599d33eda626ec5cf1280ed61ceb934ff34ab31"} Nov 22 09:44:50 crc kubenswrapper[4735]: I1122 09:44:50.193866 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6d34f31784d360c04f48988d599d33eda626ec5cf1280ed61ceb934ff34ab31" Nov 22 09:44:50 crc kubenswrapper[4735]: I1122 09:44:50.193948 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-hj44k" Nov 22 09:44:50 crc kubenswrapper[4735]: I1122 09:44:50.657062 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xgm58/crc-debug-hj44k"] Nov 22 09:44:50 crc kubenswrapper[4735]: I1122 09:44:50.670119 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xgm58/crc-debug-hj44k"] Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.333117 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70281419-fcb7-434d-bc93-eb3cc918f1da" path="/var/lib/kubelet/pods/70281419-fcb7-434d-bc93-eb3cc918f1da/volumes" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.871371 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xgm58/crc-debug-c247q"] Nov 22 09:44:51 crc kubenswrapper[4735]: E1122 09:44:51.872142 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70281419-fcb7-434d-bc93-eb3cc918f1da" containerName="container-00" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.872166 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="70281419-fcb7-434d-bc93-eb3cc918f1da" containerName="container-00" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.872709 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="70281419-fcb7-434d-bc93-eb3cc918f1da" containerName="container-00" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.874040 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.876669 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xgm58"/"default-dockercfg-tfbj6" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.885041 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtf2v\" (UniqueName: \"kubernetes.io/projected/32b56e63-6ba1-4bc5-81a6-76ebd8445074-kube-api-access-jtf2v\") pod \"crc-debug-c247q\" (UID: \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\") " pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.885228 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32b56e63-6ba1-4bc5-81a6-76ebd8445074-host\") pod \"crc-debug-c247q\" (UID: \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\") " pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.988413 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtf2v\" (UniqueName: \"kubernetes.io/projected/32b56e63-6ba1-4bc5-81a6-76ebd8445074-kube-api-access-jtf2v\") pod \"crc-debug-c247q\" (UID: \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\") " pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.988974 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32b56e63-6ba1-4bc5-81a6-76ebd8445074-host\") pod \"crc-debug-c247q\" (UID: \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\") " pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:51 crc kubenswrapper[4735]: I1122 09:44:51.989237 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32b56e63-6ba1-4bc5-81a6-76ebd8445074-host\") pod \"crc-debug-c247q\" (UID: \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\") " pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:52 crc kubenswrapper[4735]: I1122 09:44:52.009115 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtf2v\" (UniqueName: \"kubernetes.io/projected/32b56e63-6ba1-4bc5-81a6-76ebd8445074-kube-api-access-jtf2v\") pod \"crc-debug-c247q\" (UID: \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\") " pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:52 crc kubenswrapper[4735]: I1122 09:44:52.197304 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:53 crc kubenswrapper[4735]: I1122 09:44:53.230239 4735 generic.go:334] "Generic (PLEG): container finished" podID="32b56e63-6ba1-4bc5-81a6-76ebd8445074" containerID="e07f676fb9ed965488cc541dbb47f298b9a248c5fd204bf66cc0fb3160217e0a" exitCode=0 Nov 22 09:44:53 crc kubenswrapper[4735]: I1122 09:44:53.230377 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/crc-debug-c247q" event={"ID":"32b56e63-6ba1-4bc5-81a6-76ebd8445074","Type":"ContainerDied","Data":"e07f676fb9ed965488cc541dbb47f298b9a248c5fd204bf66cc0fb3160217e0a"} Nov 22 09:44:53 crc kubenswrapper[4735]: I1122 09:44:53.230842 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/crc-debug-c247q" event={"ID":"32b56e63-6ba1-4bc5-81a6-76ebd8445074","Type":"ContainerStarted","Data":"e81fffbf2cb1074b0b5a932ca76b62217b347124188cb24c01ce5558066315af"} Nov 22 09:44:53 crc kubenswrapper[4735]: I1122 09:44:53.283440 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xgm58/crc-debug-c247q"] Nov 22 09:44:53 crc kubenswrapper[4735]: I1122 09:44:53.296525 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xgm58/crc-debug-c247q"] Nov 22 09:44:54 crc kubenswrapper[4735]: I1122 09:44:54.395260 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:54 crc kubenswrapper[4735]: I1122 09:44:54.554035 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtf2v\" (UniqueName: \"kubernetes.io/projected/32b56e63-6ba1-4bc5-81a6-76ebd8445074-kube-api-access-jtf2v\") pod \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\" (UID: \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\") " Nov 22 09:44:54 crc kubenswrapper[4735]: I1122 09:44:54.554156 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32b56e63-6ba1-4bc5-81a6-76ebd8445074-host\") pod \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\" (UID: \"32b56e63-6ba1-4bc5-81a6-76ebd8445074\") " Nov 22 09:44:54 crc kubenswrapper[4735]: I1122 09:44:54.554248 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32b56e63-6ba1-4bc5-81a6-76ebd8445074-host" (OuterVolumeSpecName: "host") pod "32b56e63-6ba1-4bc5-81a6-76ebd8445074" (UID: "32b56e63-6ba1-4bc5-81a6-76ebd8445074"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 09:44:54 crc kubenswrapper[4735]: I1122 09:44:54.555059 4735 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32b56e63-6ba1-4bc5-81a6-76ebd8445074-host\") on node \"crc\" DevicePath \"\"" Nov 22 09:44:54 crc kubenswrapper[4735]: I1122 09:44:54.559276 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b56e63-6ba1-4bc5-81a6-76ebd8445074-kube-api-access-jtf2v" (OuterVolumeSpecName: "kube-api-access-jtf2v") pod "32b56e63-6ba1-4bc5-81a6-76ebd8445074" (UID: "32b56e63-6ba1-4bc5-81a6-76ebd8445074"). InnerVolumeSpecName "kube-api-access-jtf2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:44:54 crc kubenswrapper[4735]: I1122 09:44:54.657340 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtf2v\" (UniqueName: \"kubernetes.io/projected/32b56e63-6ba1-4bc5-81a6-76ebd8445074-kube-api-access-jtf2v\") on node \"crc\" DevicePath \"\"" Nov 22 09:44:55 crc kubenswrapper[4735]: I1122 09:44:55.254835 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e81fffbf2cb1074b0b5a932ca76b62217b347124188cb24c01ce5558066315af" Nov 22 09:44:55 crc kubenswrapper[4735]: I1122 09:44:55.254875 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/crc-debug-c247q" Nov 22 09:44:55 crc kubenswrapper[4735]: I1122 09:44:55.279291 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b56e63-6ba1-4bc5-81a6-76ebd8445074" path="/var/lib/kubelet/pods/32b56e63-6ba1-4bc5-81a6-76ebd8445074/volumes" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.190939 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs"] Nov 22 09:45:00 crc kubenswrapper[4735]: E1122 09:45:00.192076 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b56e63-6ba1-4bc5-81a6-76ebd8445074" containerName="container-00" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.192095 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b56e63-6ba1-4bc5-81a6-76ebd8445074" containerName="container-00" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.192478 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b56e63-6ba1-4bc5-81a6-76ebd8445074" containerName="container-00" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.193497 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.196288 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpz9n\" (UniqueName: \"kubernetes.io/projected/d946af89-dec5-4113-a3d0-93ab141c8e92-kube-api-access-vpz9n\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.196430 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d946af89-dec5-4113-a3d0-93ab141c8e92-secret-volume\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.196608 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d946af89-dec5-4113-a3d0-93ab141c8e92-config-volume\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.198937 4735 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.199142 4735 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.201298 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs"] Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.298546 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d946af89-dec5-4113-a3d0-93ab141c8e92-secret-volume\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.298639 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d946af89-dec5-4113-a3d0-93ab141c8e92-config-volume\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.298948 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpz9n\" (UniqueName: \"kubernetes.io/projected/d946af89-dec5-4113-a3d0-93ab141c8e92-kube-api-access-vpz9n\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.300144 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d946af89-dec5-4113-a3d0-93ab141c8e92-config-volume\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.311607 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d946af89-dec5-4113-a3d0-93ab141c8e92-secret-volume\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.324122 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpz9n\" (UniqueName: \"kubernetes.io/projected/d946af89-dec5-4113-a3d0-93ab141c8e92-kube-api-access-vpz9n\") pod \"collect-profiles-29396745-5f7cs\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:00 crc kubenswrapper[4735]: I1122 09:45:00.517970 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:01 crc kubenswrapper[4735]: W1122 09:45:01.370599 4735 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd946af89_dec5_4113_a3d0_93ab141c8e92.slice/crio-c8353f58e7da0ad8c3237f391169c8d9f3528a94479e1d64b02735358a96b935 WatchSource:0}: Error finding container c8353f58e7da0ad8c3237f391169c8d9f3528a94479e1d64b02735358a96b935: Status 404 returned error can't find the container with id c8353f58e7da0ad8c3237f391169c8d9f3528a94479e1d64b02735358a96b935 Nov 22 09:45:01 crc kubenswrapper[4735]: I1122 09:45:01.371929 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs"] Nov 22 09:45:02 crc kubenswrapper[4735]: I1122 09:45:02.340609 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" event={"ID":"d946af89-dec5-4113-a3d0-93ab141c8e92","Type":"ContainerStarted","Data":"a2810b0efa91e3510c901d5f55893376eee24160ad20315a3f30ada1f987fa24"} Nov 22 09:45:02 crc kubenswrapper[4735]: I1122 09:45:02.340900 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" event={"ID":"d946af89-dec5-4113-a3d0-93ab141c8e92","Type":"ContainerStarted","Data":"c8353f58e7da0ad8c3237f391169c8d9f3528a94479e1d64b02735358a96b935"} Nov 22 09:45:02 crc kubenswrapper[4735]: I1122 09:45:02.363661 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" podStartSLOduration=2.363633745 podStartE2EDuration="2.363633745s" podCreationTimestamp="2025-11-22 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 09:45:02.357122086 +0000 UTC m=+6123.961460701" watchObservedRunningTime="2025-11-22 09:45:02.363633745 +0000 UTC m=+6123.967972350" Nov 22 09:45:03 crc kubenswrapper[4735]: I1122 09:45:03.352940 4735 generic.go:334] "Generic (PLEG): container finished" podID="d946af89-dec5-4113-a3d0-93ab141c8e92" containerID="a2810b0efa91e3510c901d5f55893376eee24160ad20315a3f30ada1f987fa24" exitCode=0 Nov 22 09:45:03 crc kubenswrapper[4735]: I1122 09:45:03.352996 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" event={"ID":"d946af89-dec5-4113-a3d0-93ab141c8e92","Type":"ContainerDied","Data":"a2810b0efa91e3510c901d5f55893376eee24160ad20315a3f30ada1f987fa24"} Nov 22 09:45:04 crc kubenswrapper[4735]: I1122 09:45:04.761025 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:04 crc kubenswrapper[4735]: I1122 09:45:04.907063 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d946af89-dec5-4113-a3d0-93ab141c8e92-config-volume\") pod \"d946af89-dec5-4113-a3d0-93ab141c8e92\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " Nov 22 09:45:04 crc kubenswrapper[4735]: I1122 09:45:04.907227 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d946af89-dec5-4113-a3d0-93ab141c8e92-secret-volume\") pod \"d946af89-dec5-4113-a3d0-93ab141c8e92\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " Nov 22 09:45:04 crc kubenswrapper[4735]: I1122 09:45:04.907319 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpz9n\" (UniqueName: \"kubernetes.io/projected/d946af89-dec5-4113-a3d0-93ab141c8e92-kube-api-access-vpz9n\") pod \"d946af89-dec5-4113-a3d0-93ab141c8e92\" (UID: \"d946af89-dec5-4113-a3d0-93ab141c8e92\") " Nov 22 09:45:04 crc kubenswrapper[4735]: I1122 09:45:04.907760 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d946af89-dec5-4113-a3d0-93ab141c8e92-config-volume" (OuterVolumeSpecName: "config-volume") pod "d946af89-dec5-4113-a3d0-93ab141c8e92" (UID: "d946af89-dec5-4113-a3d0-93ab141c8e92"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:45:04 crc kubenswrapper[4735]: I1122 09:45:04.908206 4735 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d946af89-dec5-4113-a3d0-93ab141c8e92-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:04 crc kubenswrapper[4735]: I1122 09:45:04.913851 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d946af89-dec5-4113-a3d0-93ab141c8e92-kube-api-access-vpz9n" (OuterVolumeSpecName: "kube-api-access-vpz9n") pod "d946af89-dec5-4113-a3d0-93ab141c8e92" (UID: "d946af89-dec5-4113-a3d0-93ab141c8e92"). InnerVolumeSpecName "kube-api-access-vpz9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:45:04 crc kubenswrapper[4735]: I1122 09:45:04.914271 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d946af89-dec5-4113-a3d0-93ab141c8e92-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d946af89-dec5-4113-a3d0-93ab141c8e92" (UID: "d946af89-dec5-4113-a3d0-93ab141c8e92"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:45:05 crc kubenswrapper[4735]: I1122 09:45:05.010141 4735 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d946af89-dec5-4113-a3d0-93ab141c8e92-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:05 crc kubenswrapper[4735]: I1122 09:45:05.010173 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpz9n\" (UniqueName: \"kubernetes.io/projected/d946af89-dec5-4113-a3d0-93ab141c8e92-kube-api-access-vpz9n\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:05 crc kubenswrapper[4735]: I1122 09:45:05.374957 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" event={"ID":"d946af89-dec5-4113-a3d0-93ab141c8e92","Type":"ContainerDied","Data":"c8353f58e7da0ad8c3237f391169c8d9f3528a94479e1d64b02735358a96b935"} Nov 22 09:45:05 crc kubenswrapper[4735]: I1122 09:45:05.375000 4735 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8353f58e7da0ad8c3237f391169c8d9f3528a94479e1d64b02735358a96b935" Nov 22 09:45:05 crc kubenswrapper[4735]: I1122 09:45:05.375062 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-5f7cs" Nov 22 09:45:05 crc kubenswrapper[4735]: I1122 09:45:05.439723 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p"] Nov 22 09:45:05 crc kubenswrapper[4735]: I1122 09:45:05.458385 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-znr6p"] Nov 22 09:45:07 crc kubenswrapper[4735]: I1122 09:45:07.296241 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf0172f-b54d-493e-8e69-aaae5d05459c" path="/var/lib/kubelet/pods/eaf0172f-b54d-493e-8e69-aaae5d05459c/volumes" Nov 22 09:45:12 crc kubenswrapper[4735]: I1122 09:45:12.982447 4735 scope.go:117] "RemoveContainer" containerID="e1199b3e5316f5bd1beed2d3ca40bce6c0a7d52ad048a1a24c9b7d9266efad54" Nov 22 09:45:16 crc kubenswrapper[4735]: I1122 09:45:16.131910 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:45:16 crc kubenswrapper[4735]: I1122 09:45:16.132752 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.003314 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d211c51b-4464-4109-9edf-3b05feddb072/aodh-api/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.172055 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d211c51b-4464-4109-9edf-3b05feddb072/aodh-evaluator/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.192485 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d211c51b-4464-4109-9edf-3b05feddb072/aodh-listener/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.215390 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d211c51b-4464-4109-9edf-3b05feddb072/aodh-notifier/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.381722 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-cdffccfd-kxg89_9993180a-9f49-4fce-a831-ca0b4ef58428/barbican-api/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.390210 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-cdffccfd-kxg89_9993180a-9f49-4fce-a831-ca0b4ef58428/barbican-api-log/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.517285 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-66b55c68fd-lfj8p_736e8f87-4b05-4335-954a-47d3c89f84fc/barbican-keystone-listener/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.733960 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-66b55c68fd-lfj8p_736e8f87-4b05-4335-954a-47d3c89f84fc/barbican-keystone-listener-log/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.867891 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69b9b656dc-jwxhg_0966ff10-5a13-45a4-a20a-5eb889270c5b/barbican-worker/0.log" Nov 22 09:45:20 crc kubenswrapper[4735]: I1122 09:45:20.957637 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69b9b656dc-jwxhg_0966ff10-5a13-45a4-a20a-5eb889270c5b/barbican-worker-log/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.062691 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-r7sfr_d5c15b81-20c1-46c8-8f80-dcb4e7394735/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.232862 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_be212b2c-6753-40d9-9bee-0dfe2ff10dbc/ceilometer-central-agent/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.282693 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_be212b2c-6753-40d9-9bee-0dfe2ff10dbc/ceilometer-notification-agent/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.311128 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_be212b2c-6753-40d9-9bee-0dfe2ff10dbc/proxy-httpd/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.360785 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_be212b2c-6753-40d9-9bee-0dfe2ff10dbc/sg-core/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.501540 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5494e4ea-bbac-4c48-ac26-7141cc543bb4/cinder-api-log/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.560408 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5494e4ea-bbac-4c48-ac26-7141cc543bb4/cinder-api/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.746480 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_55ef656f-16ef-4187-9d57-01ab16e11c17/cinder-scheduler/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.856439 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4vbw7_ab4508bb-701a-4c44-bdea-5f3e176afe47/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:21 crc kubenswrapper[4735]: I1122 09:45:21.859348 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_55ef656f-16ef-4187-9d57-01ab16e11c17/probe/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.057253 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-dh7nx_ce09bb64-050b-47f7-b2dc-2673c5444bf0/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.118625 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-8x2j8_61b5bedf-84ff-4392-9435-0f53415cc123/init/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.288063 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-8x2j8_61b5bedf-84ff-4392-9435-0f53415cc123/init/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.348930 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-8x2j8_61b5bedf-84ff-4392-9435-0f53415cc123/dnsmasq-dns/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.438927 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-kkbcw_8b249f3f-b08f-44b7-8652-9db500ef58b5/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.535233 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2d252eab-255d-49ac-a06a-de9665af67d4/glance-httpd/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.555955 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2d252eab-255d-49ac-a06a-de9665af67d4/glance-log/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.771821 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b9513901-cb1d-4603-9c3f-7fcc2e96f162/glance-httpd/0.log" Nov 22 09:45:22 crc kubenswrapper[4735]: I1122 09:45:22.774946 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b9513901-cb1d-4603-9c3f-7fcc2e96f162/glance-log/0.log" Nov 22 09:45:23 crc kubenswrapper[4735]: I1122 09:45:23.424616 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7fc4dcbc59-gp5ct_6a03b596-b9b1-4efd-b55d-711cb33f7f78/heat-engine/0.log" Nov 22 09:45:23 crc kubenswrapper[4735]: I1122 09:45:23.725157 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-kth9z_506cbc0b-4b85-42e9-a73a-52721f001c05/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:23 crc kubenswrapper[4735]: I1122 09:45:23.736621 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-68bf475768-f4hbn_7eaaafaf-ea3f-4ba9-b558-e128f1fdde64/heat-api/0.log" Nov 22 09:45:23 crc kubenswrapper[4735]: I1122 09:45:23.830132 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-m5dfm_a8e6f487-27cc-4062-8ccc-ba2a7f06454d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:23 crc kubenswrapper[4735]: I1122 09:45:23.838844 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5d98d484bf-d5s85_a4486150-d780-4df9-a6a4-9aa0b78c468f/heat-cfnapi/0.log" Nov 22 09:45:24 crc kubenswrapper[4735]: I1122 09:45:24.074123 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29396701-zh6s9_0d2bda54-d543-40c3-8cfd-fb9e03080130/keystone-cron/0.log" Nov 22 09:45:24 crc kubenswrapper[4735]: I1122 09:45:24.388227 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5584fd997b-7ztjv_4906aa84-20dc-4096-8d7c-0ed5dd4d9103/keystone-api/0.log" Nov 22 09:45:24 crc kubenswrapper[4735]: I1122 09:45:24.389102 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_378d8516-81a6-4d5a-99cd-d5c4190370af/kube-state-metrics/0.log" Nov 22 09:45:24 crc kubenswrapper[4735]: I1122 09:45:24.478753 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-mpl9w_341a140a-5530-434a-b245-e07c4a7068fc/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:24 crc kubenswrapper[4735]: I1122 09:45:24.688358 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-pnpt5_3a770eb5-beff-422b-ac8d-a98899d08a49/logging-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:24 crc kubenswrapper[4735]: I1122 09:45:24.988128 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_da023172-a0d0-4420-be51-05b348cf6c0c/mysqld-exporter/0.log" Nov 22 09:45:25 crc kubenswrapper[4735]: I1122 09:45:25.316726 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-4gcv2_06f78f85-aa4b-41bc-8b12-3fefeca970ee/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:25 crc kubenswrapper[4735]: I1122 09:45:25.338544 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7bb54fbc4c-fnr9g_007a6147-937e-485d-9e4c-1a193fdff463/neutron-httpd/0.log" Nov 22 09:45:25 crc kubenswrapper[4735]: I1122 09:45:25.378841 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7bb54fbc4c-fnr9g_007a6147-937e-485d-9e4c-1a193fdff463/neutron-api/0.log" Nov 22 09:45:26 crc kubenswrapper[4735]: I1122 09:45:26.000133 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7ffa0c54-0cf4-431a-932a-29cc50706cf1/nova-cell0-conductor-conductor/0.log" Nov 22 09:45:26 crc kubenswrapper[4735]: I1122 09:45:26.194378 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_14026d8e-f176-4c9f-8092-104e9a59ed5c/nova-api-log/0.log" Nov 22 09:45:26 crc kubenswrapper[4735]: I1122 09:45:26.230727 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b04e0725-3b2f-411c-aeaa-1c9d586bbe37/nova-cell1-conductor-conductor/0.log" Nov 22 09:45:26 crc kubenswrapper[4735]: I1122 09:45:26.546768 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-cf6ns_1867df3c-9d3b-41b0-9d6f-153536f35dd8/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:26 crc kubenswrapper[4735]: I1122 09:45:26.612544 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_939a7da6-6cb9-4e53-8875-357317b90738/nova-cell1-novncproxy-novncproxy/0.log" Nov 22 09:45:26 crc kubenswrapper[4735]: I1122 09:45:26.657855 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_14026d8e-f176-4c9f-8092-104e9a59ed5c/nova-api-api/0.log" Nov 22 09:45:26 crc kubenswrapper[4735]: I1122 09:45:26.898294 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c346a589-fa62-4d07-9fcb-c4afb7f52b45/nova-metadata-log/0.log" Nov 22 09:45:27 crc kubenswrapper[4735]: I1122 09:45:27.047773 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a47abba7-dce0-433f-a78e-1b3224e07835/nova-scheduler-scheduler/0.log" Nov 22 09:45:27 crc kubenswrapper[4735]: I1122 09:45:27.131919 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0c715895-af70-4e8f-96c4-c7064e45c2c9/mysql-bootstrap/0.log" Nov 22 09:45:27 crc kubenswrapper[4735]: I1122 09:45:27.363796 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0c715895-af70-4e8f-96c4-c7064e45c2c9/galera/0.log" Nov 22 09:45:27 crc kubenswrapper[4735]: I1122 09:45:27.375813 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0c715895-af70-4e8f-96c4-c7064e45c2c9/mysql-bootstrap/0.log" Nov 22 09:45:27 crc kubenswrapper[4735]: I1122 09:45:27.580587 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5eaebfb3-8113-41d6-9a3d-77c8194ad4d1/mysql-bootstrap/0.log" Nov 22 09:45:27 crc kubenswrapper[4735]: I1122 09:45:27.762149 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5eaebfb3-8113-41d6-9a3d-77c8194ad4d1/mysql-bootstrap/0.log" Nov 22 09:45:27 crc kubenswrapper[4735]: I1122 09:45:27.803744 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5eaebfb3-8113-41d6-9a3d-77c8194ad4d1/galera/0.log" Nov 22 09:45:27 crc kubenswrapper[4735]: I1122 09:45:27.996295 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_da8872ad-7c19-451d-84be-6333126bfb63/openstackclient/0.log" Nov 22 09:45:28 crc kubenswrapper[4735]: I1122 09:45:28.108678 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4tx8r_451e7b5a-64f4-4f0a-8a44-32a86375b765/openstack-network-exporter/0.log" Nov 22 09:45:28 crc kubenswrapper[4735]: I1122 09:45:28.467128 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7mhwj_fee31945-9fca-4693-9218-97bb4513cee3/ovsdb-server-init/0.log" Nov 22 09:45:28 crc kubenswrapper[4735]: I1122 09:45:28.552754 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7mhwj_fee31945-9fca-4693-9218-97bb4513cee3/ovs-vswitchd/0.log" Nov 22 09:45:28 crc kubenswrapper[4735]: I1122 09:45:28.558367 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7mhwj_fee31945-9fca-4693-9218-97bb4513cee3/ovsdb-server-init/0.log" Nov 22 09:45:28 crc kubenswrapper[4735]: I1122 09:45:28.714737 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7mhwj_fee31945-9fca-4693-9218-97bb4513cee3/ovsdb-server/0.log" Nov 22 09:45:28 crc kubenswrapper[4735]: I1122 09:45:28.858796 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z5csg_d99e6f50-a748-4ae8-b1b0-9df901b37201/ovn-controller/0.log" Nov 22 09:45:29 crc kubenswrapper[4735]: I1122 09:45:29.052991 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c346a589-fa62-4d07-9fcb-c4afb7f52b45/nova-metadata-metadata/0.log" Nov 22 09:45:29 crc kubenswrapper[4735]: I1122 09:45:29.080506 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-lgnq6_311958ee-5838-4326-9b40-7b7228bd6393/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:29 crc kubenswrapper[4735]: I1122 09:45:29.308702 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3ebf2f32-a55d-4844-85b3-bb09482d04e1/openstack-network-exporter/0.log" Nov 22 09:45:29 crc kubenswrapper[4735]: I1122 09:45:29.327518 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3ebf2f32-a55d-4844-85b3-bb09482d04e1/ovn-northd/0.log" Nov 22 09:45:29 crc kubenswrapper[4735]: I1122 09:45:29.398359 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_16a0218b-2a6b-47f5-832f-2d8da9f950f3/openstack-network-exporter/0.log" Nov 22 09:45:29 crc kubenswrapper[4735]: I1122 09:45:29.549509 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_16a0218b-2a6b-47f5-832f-2d8da9f950f3/ovsdbserver-nb/0.log" Nov 22 09:45:29 crc kubenswrapper[4735]: I1122 09:45:29.591745 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_135fc91d-a8bd-489a-a7ae-3df9db9af6cd/openstack-network-exporter/0.log" Nov 22 09:45:29 crc kubenswrapper[4735]: I1122 09:45:29.709652 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_135fc91d-a8bd-489a-a7ae-3df9db9af6cd/ovsdbserver-sb/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.012906 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f1d9efee-4efc-460c-ba99-36df95d96a48/init-config-reloader/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.021282 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6fcc565756-vf564_99fbc92e-2676-41c5-abe1-9fc69aa0dad3/placement-api/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.098299 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6fcc565756-vf564_99fbc92e-2676-41c5-abe1-9fc69aa0dad3/placement-log/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.269861 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f1d9efee-4efc-460c-ba99-36df95d96a48/init-config-reloader/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.280759 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f1d9efee-4efc-460c-ba99-36df95d96a48/config-reloader/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.286247 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f1d9efee-4efc-460c-ba99-36df95d96a48/prometheus/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.310136 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f1d9efee-4efc-460c-ba99-36df95d96a48/thanos-sidecar/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.461433 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_019d2c19-2804-4eb6-a150-8dc1163f28c6/setup-container/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.703744 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_019d2c19-2804-4eb6-a150-8dc1163f28c6/rabbitmq/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.767188 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ac149374-df2c-4533-b55e-071b860e88af/setup-container/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.787559 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_019d2c19-2804-4eb6-a150-8dc1163f28c6/setup-container/0.log" Nov 22 09:45:30 crc kubenswrapper[4735]: I1122 09:45:30.992843 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ac149374-df2c-4533-b55e-071b860e88af/setup-container/0.log" Nov 22 09:45:31 crc kubenswrapper[4735]: I1122 09:45:31.043723 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-tf8nw_99efb16a-02eb-434a-9928-7c782fd7fc86/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:31 crc kubenswrapper[4735]: I1122 09:45:31.048417 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ac149374-df2c-4533-b55e-071b860e88af/rabbitmq/0.log" Nov 22 09:45:31 crc kubenswrapper[4735]: I1122 09:45:31.303533 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2wqk7_de67d031-9e4d-4ef8-bb18-f4d9fb04835c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:31 crc kubenswrapper[4735]: I1122 09:45:31.310633 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-xkqtx_dc4a5c09-e30f-4ab9-8668-52fbd81751c5/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:31 crc kubenswrapper[4735]: I1122 09:45:31.543842 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-6vlw6_179b352a-82a0-438e-9e6d-79fa7b20ac90/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:31 crc kubenswrapper[4735]: I1122 09:45:31.580889 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-km7xk_8eac2dec-6b49-43a8-a603-60272938cbf4/ssh-known-hosts-edpm-deployment/0.log" Nov 22 09:45:31 crc kubenswrapper[4735]: I1122 09:45:31.771193 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-d55fd4b4f-24s49_b3fc589b-feac-438c-b008-8911417ccfa7/proxy-server/0.log" Nov 22 09:45:31 crc kubenswrapper[4735]: I1122 09:45:31.991095 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-d55fd4b4f-24s49_b3fc589b-feac-438c-b008-8911417ccfa7/proxy-httpd/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.178636 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-hvj22_abf32873-0945-47b6-8f4f-ed1b18dc6eb1/swift-ring-rebalance/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.226285 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/account-auditor/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.275727 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/account-reaper/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.455128 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/account-replicator/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.518431 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/account-server/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.545164 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/container-replicator/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.550261 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/container-auditor/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.709273 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/container-server/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.769404 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/container-updater/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.853842 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/object-auditor/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.870250 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/object-expirer/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.964540 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/object-replicator/0.log" Nov 22 09:45:32 crc kubenswrapper[4735]: I1122 09:45:32.995546 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/object-server/0.log" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.076761 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/object-updater/0.log" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.101626 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/rsync/0.log" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.177424 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_49e1ddee-0d04-4d19-9c87-e438e5e70e70/swift-recon-cron/0.log" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.340767 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-tgvgl_4d1a66ad-914a-4d94-ba26-7ed039cadf8f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.459397 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-cjl2r_a393bb4a-7413-4a36-bc56-e6a18f17fffa/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.579482 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vtrk6"] Nov 22 09:45:33 crc kubenswrapper[4735]: E1122 09:45:33.580163 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d946af89-dec5-4113-a3d0-93ab141c8e92" containerName="collect-profiles" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.580186 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="d946af89-dec5-4113-a3d0-93ab141c8e92" containerName="collect-profiles" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.580586 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="d946af89-dec5-4113-a3d0-93ab141c8e92" containerName="collect-profiles" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.582804 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.590422 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vtrk6"] Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.719136 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-catalog-content\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.719250 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkh6w\" (UniqueName: \"kubernetes.io/projected/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-kube-api-access-nkh6w\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.719500 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-utilities\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.732084 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c68fad70-fe37-44d9-9fce-83f3e4c5d55a/test-operator-logs-container/0.log" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.821017 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-utilities\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.821101 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-catalog-content\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.821153 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkh6w\" (UniqueName: \"kubernetes.io/projected/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-kube-api-access-nkh6w\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.821871 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-catalog-content\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.823685 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-utilities\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.842300 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkh6w\" (UniqueName: \"kubernetes.io/projected/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-kube-api-access-nkh6w\") pod \"community-operators-vtrk6\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.884430 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-76smf_00a5290b-dff8-438f-a047-33f1d5f05f4d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 09:45:33 crc kubenswrapper[4735]: I1122 09:45:33.919335 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:34 crc kubenswrapper[4735]: I1122 09:45:34.481254 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vtrk6"] Nov 22 09:45:34 crc kubenswrapper[4735]: I1122 09:45:34.739378 4735 generic.go:334] "Generic (PLEG): container finished" podID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerID="80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837" exitCode=0 Nov 22 09:45:34 crc kubenswrapper[4735]: I1122 09:45:34.739500 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtrk6" event={"ID":"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc","Type":"ContainerDied","Data":"80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837"} Nov 22 09:45:34 crc kubenswrapper[4735]: I1122 09:45:34.744886 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtrk6" event={"ID":"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc","Type":"ContainerStarted","Data":"7bca3bb44871f526cdb3bbad8e38ea7b52fe8b17d8eb5af82488e3a59ad965da"} Nov 22 09:45:34 crc kubenswrapper[4735]: I1122 09:45:34.831329 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_9d89aa19-7bdb-4db2-ad64-438da4087c36/tempest-tests-tempest-tests-runner/0.log" Nov 22 09:45:36 crc kubenswrapper[4735]: I1122 09:45:36.772404 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtrk6" event={"ID":"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc","Type":"ContainerStarted","Data":"54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808"} Nov 22 09:45:37 crc kubenswrapper[4735]: I1122 09:45:37.787150 4735 generic.go:334] "Generic (PLEG): container finished" podID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerID="54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808" exitCode=0 Nov 22 09:45:37 crc kubenswrapper[4735]: I1122 09:45:37.787195 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtrk6" event={"ID":"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc","Type":"ContainerDied","Data":"54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808"} Nov 22 09:45:38 crc kubenswrapper[4735]: I1122 09:45:38.800500 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtrk6" event={"ID":"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc","Type":"ContainerStarted","Data":"2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff"} Nov 22 09:45:38 crc kubenswrapper[4735]: I1122 09:45:38.828747 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vtrk6" podStartSLOduration=2.339909917 podStartE2EDuration="5.828726352s" podCreationTimestamp="2025-11-22 09:45:33 +0000 UTC" firstStartedPulling="2025-11-22 09:45:34.74259015 +0000 UTC m=+6156.346928755" lastFinishedPulling="2025-11-22 09:45:38.231406585 +0000 UTC m=+6159.835745190" observedRunningTime="2025-11-22 09:45:38.817161976 +0000 UTC m=+6160.421500581" watchObservedRunningTime="2025-11-22 09:45:38.828726352 +0000 UTC m=+6160.433064957" Nov 22 09:45:41 crc kubenswrapper[4735]: I1122 09:45:41.985979 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_b0c3423f-5a2e-4e2f-a54f-ca67d4e466ef/memcached/0.log" Nov 22 09:45:43 crc kubenswrapper[4735]: I1122 09:45:43.920515 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:43 crc kubenswrapper[4735]: I1122 09:45:43.922020 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:44 crc kubenswrapper[4735]: I1122 09:45:44.973401 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-vtrk6" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="registry-server" probeResult="failure" output=< Nov 22 09:45:44 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:45:44 crc kubenswrapper[4735]: > Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.131258 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.132234 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.132389 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.133398 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"76f202ccd08e999858681fe56c8b8a5850e9e3360e4ffa862758f705684ac588"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.133565 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://76f202ccd08e999858681fe56c8b8a5850e9e3360e4ffa862758f705684ac588" gracePeriod=600 Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.910683 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="76f202ccd08e999858681fe56c8b8a5850e9e3360e4ffa862758f705684ac588" exitCode=0 Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.910765 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"76f202ccd08e999858681fe56c8b8a5850e9e3360e4ffa862758f705684ac588"} Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.911243 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827"} Nov 22 09:45:46 crc kubenswrapper[4735]: I1122 09:45:46.911273 4735 scope.go:117] "RemoveContainer" containerID="93de38872dd7f6b8f684ed9af9bb118489aaf65866213c48d1f50a5d7cba8ff4" Nov 22 09:45:53 crc kubenswrapper[4735]: I1122 09:45:53.970898 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:54 crc kubenswrapper[4735]: I1122 09:45:54.030665 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:54 crc kubenswrapper[4735]: I1122 09:45:54.213031 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vtrk6"] Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.034176 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vtrk6" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="registry-server" containerID="cri-o://2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff" gracePeriod=2 Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.517688 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.647083 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-utilities\") pod \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.647684 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkh6w\" (UniqueName: \"kubernetes.io/projected/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-kube-api-access-nkh6w\") pod \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.647831 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-catalog-content\") pod \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\" (UID: \"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc\") " Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.648120 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-utilities" (OuterVolumeSpecName: "utilities") pod "87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" (UID: "87da5b1a-e60e-4ea0-9baa-4d285dc53ffc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.650978 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.660711 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-kube-api-access-nkh6w" (OuterVolumeSpecName: "kube-api-access-nkh6w") pod "87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" (UID: "87da5b1a-e60e-4ea0-9baa-4d285dc53ffc"). InnerVolumeSpecName "kube-api-access-nkh6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.710231 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" (UID: "87da5b1a-e60e-4ea0-9baa-4d285dc53ffc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.752830 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkh6w\" (UniqueName: \"kubernetes.io/projected/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-kube-api-access-nkh6w\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:55 crc kubenswrapper[4735]: I1122 09:45:55.752866 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.046880 4735 generic.go:334] "Generic (PLEG): container finished" podID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerID="2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff" exitCode=0 Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.046937 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtrk6" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.046956 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtrk6" event={"ID":"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc","Type":"ContainerDied","Data":"2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff"} Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.047401 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtrk6" event={"ID":"87da5b1a-e60e-4ea0-9baa-4d285dc53ffc","Type":"ContainerDied","Data":"7bca3bb44871f526cdb3bbad8e38ea7b52fe8b17d8eb5af82488e3a59ad965da"} Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.047434 4735 scope.go:117] "RemoveContainer" containerID="2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.077155 4735 scope.go:117] "RemoveContainer" containerID="54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.091590 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vtrk6"] Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.101827 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vtrk6"] Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.118177 4735 scope.go:117] "RemoveContainer" containerID="80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.178434 4735 scope.go:117] "RemoveContainer" containerID="2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff" Nov 22 09:45:56 crc kubenswrapper[4735]: E1122 09:45:56.178998 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff\": container with ID starting with 2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff not found: ID does not exist" containerID="2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.179062 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff"} err="failed to get container status \"2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff\": rpc error: code = NotFound desc = could not find container \"2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff\": container with ID starting with 2cf77eaeb11ceadffd9c18d67901b657e9ebe0af5a0a96eb930fef6a4ff7b5ff not found: ID does not exist" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.179090 4735 scope.go:117] "RemoveContainer" containerID="54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808" Nov 22 09:45:56 crc kubenswrapper[4735]: E1122 09:45:56.179418 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808\": container with ID starting with 54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808 not found: ID does not exist" containerID="54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.179461 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808"} err="failed to get container status \"54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808\": rpc error: code = NotFound desc = could not find container \"54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808\": container with ID starting with 54485e4088978ac7860118d1f7c90ac7210f45c53bf003ef138fb9a495011808 not found: ID does not exist" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.179508 4735 scope.go:117] "RemoveContainer" containerID="80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837" Nov 22 09:45:56 crc kubenswrapper[4735]: E1122 09:45:56.179720 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837\": container with ID starting with 80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837 not found: ID does not exist" containerID="80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837" Nov 22 09:45:56 crc kubenswrapper[4735]: I1122 09:45:56.179744 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837"} err="failed to get container status \"80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837\": rpc error: code = NotFound desc = could not find container \"80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837\": container with ID starting with 80551f4f875daa3df9aecfde64a86de3a6e7476a48e4fd68ac781ea0a35a2837 not found: ID does not exist" Nov 22 09:45:57 crc kubenswrapper[4735]: I1122 09:45:57.282687 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" path="/var/lib/kubelet/pods/87da5b1a-e60e-4ea0-9baa-4d285dc53ffc/volumes" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.060601 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf_783be8f3-61ed-4470-a034-f680ec16dbe1/util/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.222773 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf_783be8f3-61ed-4470-a034-f680ec16dbe1/util/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.236258 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf_783be8f3-61ed-4470-a034-f680ec16dbe1/pull/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.304109 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf_783be8f3-61ed-4470-a034-f680ec16dbe1/pull/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.495399 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf_783be8f3-61ed-4470-a034-f680ec16dbe1/util/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.496129 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf_783be8f3-61ed-4470-a034-f680ec16dbe1/pull/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.501495 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2ba798c1e457009c84908b01c4d64b4c3200b85d6f0af3006407b3bceb8splf_783be8f3-61ed-4470-a034-f680ec16dbe1/extract/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.680144 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-dmhxb_5765b990-2731-42c7-bb14-7b149e91555f/kube-rbac-proxy/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.723474 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-dmhxb_5765b990-2731-42c7-bb14-7b149e91555f/manager/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.755803 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-26vdp_2dd266b7-8f4d-46b3-b8e7-3511b5404f8d/kube-rbac-proxy/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.917367 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-26vdp_2dd266b7-8f4d-46b3-b8e7-3511b5404f8d/manager/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.934408 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-j5xsn_2c87c2e3-b68e-4f75-a7cf-2c752e7286e8/manager/0.log" Nov 22 09:46:00 crc kubenswrapper[4735]: I1122 09:46:00.945099 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-j5xsn_2c87c2e3-b68e-4f75-a7cf-2c752e7286e8/kube-rbac-proxy/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.120289 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-zrv98_2038726f-ef9a-47d2-837e-1b81f72c2d39/kube-rbac-proxy/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.210848 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-zrv98_2038726f-ef9a-47d2-837e-1b81f72c2d39/manager/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.316979 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-gkpjb_1c5cb76f-8964-4245-86e5-56b619e06d1b/kube-rbac-proxy/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.423118 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-42kvc_c3ac6a85-2f1e-4906-ba7d-c9415f523fa8/kube-rbac-proxy/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.445323 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-gkpjb_1c5cb76f-8964-4245-86e5-56b619e06d1b/manager/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.521560 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-42kvc_c3ac6a85-2f1e-4906-ba7d-c9415f523fa8/manager/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.614987 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7875d8bb94-tqj8q_10f9fa68-fe17-4af9-8e47-fd534c8a4446/kube-rbac-proxy/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.789637 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7875d8bb94-tqj8q_10f9fa68-fe17-4af9-8e47-fd534c8a4446/manager/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.837255 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-pwlw2_61871ccb-bdb6-4222-88d9-d34c22a5797c/kube-rbac-proxy/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.856049 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-pwlw2_61871ccb-bdb6-4222-88d9-d34c22a5797c/manager/0.log" Nov 22 09:46:01 crc kubenswrapper[4735]: I1122 09:46:01.962991 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-t625d_2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb/kube-rbac-proxy/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.071125 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-t625d_2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb/manager/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.156909 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-nd4wh_fc7fded7-7eee-4227-8569-85d07ea4ecd6/kube-rbac-proxy/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.180296 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-nd4wh_fc7fded7-7eee-4227-8569-85d07ea4ecd6/manager/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.258004 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-vgsgb_3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c/kube-rbac-proxy/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.356050 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-vgsgb_3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c/manager/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.448873 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-xl2mv_9124e032-f3c0-46fa-bd17-26ed40bfd57e/kube-rbac-proxy/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.534672 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-xl2mv_9124e032-f3c0-46fa-bd17-26ed40bfd57e/manager/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.612818 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-b47vs_150b26f8-b5cd-4828-96cc-6e37000c24e7/kube-rbac-proxy/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.751621 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-b47vs_150b26f8-b5cd-4828-96cc-6e37000c24e7/manager/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.775898 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-2kf8g_e204b082-d231-48c6-af16-f4ee9d2d7f9c/kube-rbac-proxy/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.803539 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-2kf8g_e204b082-d231-48c6-af16-f4ee9d2d7f9c/manager/0.log" Nov 22 09:46:02 crc kubenswrapper[4735]: I1122 09:46:02.989402 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc_990c8bbe-2866-4568-8e30-1d9f016f0736/kube-rbac-proxy/0.log" Nov 22 09:46:03 crc kubenswrapper[4735]: I1122 09:46:03.000047 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-8k5dc_990c8bbe-2866-4568-8e30-1d9f016f0736/manager/0.log" Nov 22 09:46:03 crc kubenswrapper[4735]: I1122 09:46:03.204245 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-76b989774d-x7gnj_b83d5866-e478-4d10-967a-488ef0435473/kube-rbac-proxy/0.log" Nov 22 09:46:03 crc kubenswrapper[4735]: I1122 09:46:03.293683 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7475f94df-sfjx8_7ba21bad-f0f3-4e1e-add2-43b2ddfae492/kube-rbac-proxy/0.log" Nov 22 09:46:03 crc kubenswrapper[4735]: I1122 09:46:03.599758 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-24jzf_1eac1f05-c6ed-409f-802f-cb15adb08ef5/registry-server/0.log" Nov 22 09:46:03 crc kubenswrapper[4735]: I1122 09:46:03.721779 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7475f94df-sfjx8_7ba21bad-f0f3-4e1e-add2-43b2ddfae492/operator/0.log" Nov 22 09:46:03 crc kubenswrapper[4735]: I1122 09:46:03.831478 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-qs98r_85c63b21-5764-45af-98bc-47b1be4be90e/kube-rbac-proxy/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.057026 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-qs98r_85c63b21-5764-45af-98bc-47b1be4be90e/manager/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.069943 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-vndrw_2171ac01-b987-4062-9521-88c800930330/kube-rbac-proxy/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.155040 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-vndrw_2171ac01-b987-4062-9521-88c800930330/manager/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.326178 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-285tt_11f148e7-7f11-42cb-ae64-e5714bf56295/operator/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.384405 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-x7w9w_69fe7050-a836-427e-ba06-7f016e64b24b/kube-rbac-proxy/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.523670 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-x7w9w_69fe7050-a836-427e-ba06-7f016e64b24b/manager/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.546044 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7d76d46c6f-shjbx_ee3c788a-51f9-4808-996b-e854ba26a776/kube-rbac-proxy/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.731634 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-dcgnz_fb88fa45-ef38-49bc-b842-157bef4cd0e5/kube-rbac-proxy/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.811482 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-dcgnz_fb88fa45-ef38-49bc-b842-157bef4cd0e5/manager/0.log" Nov 22 09:46:04 crc kubenswrapper[4735]: I1122 09:46:04.910950 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-76b989774d-x7gnj_b83d5866-e478-4d10-967a-488ef0435473/manager/0.log" Nov 22 09:46:05 crc kubenswrapper[4735]: I1122 09:46:05.008889 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7d76d46c6f-shjbx_ee3c788a-51f9-4808-996b-e854ba26a776/manager/0.log" Nov 22 09:46:05 crc kubenswrapper[4735]: I1122 09:46:05.054641 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-cn8lg_fce6a4d4-f6fa-44bf-8528-7de867925bbb/kube-rbac-proxy/0.log" Nov 22 09:46:05 crc kubenswrapper[4735]: I1122 09:46:05.055848 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-cn8lg_fce6a4d4-f6fa-44bf-8528-7de867925bbb/manager/0.log" Nov 22 09:46:21 crc kubenswrapper[4735]: I1122 09:46:21.060284 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-gzwms_fb194fbb-af18-420c-ad19-004eba55e639/control-plane-machine-set-operator/0.log" Nov 22 09:46:21 crc kubenswrapper[4735]: I1122 09:46:21.237046 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-98dz8_510ea1b6-d5f7-42eb-b490-6f729076dcc3/machine-api-operator/0.log" Nov 22 09:46:21 crc kubenswrapper[4735]: I1122 09:46:21.291852 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-98dz8_510ea1b6-d5f7-42eb-b490-6f729076dcc3/kube-rbac-proxy/0.log" Nov 22 09:46:35 crc kubenswrapper[4735]: I1122 09:46:35.014008 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-dbnm5_cccf1779-d468-4e20-83ee-6a7c85b1d7ab/cert-manager-controller/0.log" Nov 22 09:46:35 crc kubenswrapper[4735]: I1122 09:46:35.214358 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-t6b7j_9a904776-a627-4df0-abf1-1832ba4a064c/cert-manager-cainjector/0.log" Nov 22 09:46:35 crc kubenswrapper[4735]: I1122 09:46:35.273229 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-26wft_24a36a64-01ec-4f38-b118-e1c6097e8965/cert-manager-webhook/0.log" Nov 22 09:46:48 crc kubenswrapper[4735]: I1122 09:46:48.911164 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-4pgd7_3a0c81e9-d65f-4b9e-8ea9-b1dc9793cabd/nmstate-console-plugin/0.log" Nov 22 09:46:49 crc kubenswrapper[4735]: I1122 09:46:49.108903 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-jslbs_45633081-910f-45c1-a095-21db07a7cc78/nmstate-handler/0.log" Nov 22 09:46:49 crc kubenswrapper[4735]: I1122 09:46:49.169544 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-hnkkh_c2e387de-32b1-4259-9301-0181b08ab88d/kube-rbac-proxy/0.log" Nov 22 09:46:49 crc kubenswrapper[4735]: I1122 09:46:49.191545 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-hnkkh_c2e387de-32b1-4259-9301-0181b08ab88d/nmstate-metrics/0.log" Nov 22 09:46:49 crc kubenswrapper[4735]: I1122 09:46:49.450742 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-swwps_fd488627-bc81-4723-b465-b39ff151d0d6/nmstate-webhook/0.log" Nov 22 09:46:49 crc kubenswrapper[4735]: I1122 09:46:49.459658 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-9gq6d_36088127-63b1-40be-8456-294d89283c41/nmstate-operator/0.log" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.187443 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-szfj5"] Nov 22 09:47:01 crc kubenswrapper[4735]: E1122 09:47:01.188763 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="extract-content" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.188782 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="extract-content" Nov 22 09:47:01 crc kubenswrapper[4735]: E1122 09:47:01.188838 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="extract-utilities" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.188846 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="extract-utilities" Nov 22 09:47:01 crc kubenswrapper[4735]: E1122 09:47:01.188867 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="registry-server" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.188874 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="registry-server" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.189163 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="87da5b1a-e60e-4ea0-9baa-4d285dc53ffc" containerName="registry-server" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.191301 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.224328 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-szfj5"] Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.245563 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-utilities\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.245745 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzsk8\" (UniqueName: \"kubernetes.io/projected/3b630869-ac40-452b-b1af-0e0b7c27f717-kube-api-access-jzsk8\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.245790 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-catalog-content\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.347631 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-utilities\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.347963 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzsk8\" (UniqueName: \"kubernetes.io/projected/3b630869-ac40-452b-b1af-0e0b7c27f717-kube-api-access-jzsk8\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.348017 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-catalog-content\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.348161 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-utilities\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.348690 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-catalog-content\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.372343 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzsk8\" (UniqueName: \"kubernetes.io/projected/3b630869-ac40-452b-b1af-0e0b7c27f717-kube-api-access-jzsk8\") pod \"redhat-operators-szfj5\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:01 crc kubenswrapper[4735]: I1122 09:47:01.516208 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:02 crc kubenswrapper[4735]: I1122 09:47:02.005726 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-szfj5"] Nov 22 09:47:02 crc kubenswrapper[4735]: I1122 09:47:02.838846 4735 generic.go:334] "Generic (PLEG): container finished" podID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerID="7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865" exitCode=0 Nov 22 09:47:02 crc kubenswrapper[4735]: I1122 09:47:02.839156 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szfj5" event={"ID":"3b630869-ac40-452b-b1af-0e0b7c27f717","Type":"ContainerDied","Data":"7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865"} Nov 22 09:47:02 crc kubenswrapper[4735]: I1122 09:47:02.839187 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szfj5" event={"ID":"3b630869-ac40-452b-b1af-0e0b7c27f717","Type":"ContainerStarted","Data":"05cf25a3b5309b18a92128656fa96431c5a9875db259cab4aeb29c9b3b55a6eb"} Nov 22 09:47:02 crc kubenswrapper[4735]: I1122 09:47:02.843695 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:47:03 crc kubenswrapper[4735]: I1122 09:47:03.709053 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5b4b6fd4bd-jc7r4_b0172333-c8a0-4e17-a312-d3362d0fe8da/kube-rbac-proxy/0.log" Nov 22 09:47:03 crc kubenswrapper[4735]: I1122 09:47:03.844272 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5b4b6fd4bd-jc7r4_b0172333-c8a0-4e17-a312-d3362d0fe8da/manager/0.log" Nov 22 09:47:03 crc kubenswrapper[4735]: I1122 09:47:03.853865 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szfj5" event={"ID":"3b630869-ac40-452b-b1af-0e0b7c27f717","Type":"ContainerStarted","Data":"19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf"} Nov 22 09:47:08 crc kubenswrapper[4735]: I1122 09:47:08.909522 4735 generic.go:334] "Generic (PLEG): container finished" podID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerID="19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf" exitCode=0 Nov 22 09:47:08 crc kubenswrapper[4735]: I1122 09:47:08.909614 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szfj5" event={"ID":"3b630869-ac40-452b-b1af-0e0b7c27f717","Type":"ContainerDied","Data":"19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf"} Nov 22 09:47:10 crc kubenswrapper[4735]: I1122 09:47:10.930257 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szfj5" event={"ID":"3b630869-ac40-452b-b1af-0e0b7c27f717","Type":"ContainerStarted","Data":"f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39"} Nov 22 09:47:10 crc kubenswrapper[4735]: I1122 09:47:10.970327 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-szfj5" podStartSLOduration=2.880189672 podStartE2EDuration="9.970302206s" podCreationTimestamp="2025-11-22 09:47:01 +0000 UTC" firstStartedPulling="2025-11-22 09:47:02.842865987 +0000 UTC m=+6244.447204592" lastFinishedPulling="2025-11-22 09:47:09.932978521 +0000 UTC m=+6251.537317126" observedRunningTime="2025-11-22 09:47:10.952229232 +0000 UTC m=+6252.556567847" watchObservedRunningTime="2025-11-22 09:47:10.970302206 +0000 UTC m=+6252.574640821" Nov 22 09:47:11 crc kubenswrapper[4735]: I1122 09:47:11.517089 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:11 crc kubenswrapper[4735]: I1122 09:47:11.517152 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:12 crc kubenswrapper[4735]: I1122 09:47:12.576749 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-szfj5" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="registry-server" probeResult="failure" output=< Nov 22 09:47:12 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:47:12 crc kubenswrapper[4735]: > Nov 22 09:47:13 crc kubenswrapper[4735]: I1122 09:47:13.913055 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wcpq5"] Nov 22 09:47:13 crc kubenswrapper[4735]: I1122 09:47:13.916322 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:13 crc kubenswrapper[4735]: I1122 09:47:13.933025 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wcpq5"] Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.085793 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-utilities\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.085915 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-catalog-content\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.085959 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghhb\" (UniqueName: \"kubernetes.io/projected/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-kube-api-access-rghhb\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.188765 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-catalog-content\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.188877 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghhb\" (UniqueName: \"kubernetes.io/projected/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-kube-api-access-rghhb\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.189036 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-utilities\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.189517 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-catalog-content\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.189645 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-utilities\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.222986 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghhb\" (UniqueName: \"kubernetes.io/projected/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-kube-api-access-rghhb\") pod \"certified-operators-wcpq5\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:14 crc kubenswrapper[4735]: I1122 09:47:14.245050 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:15 crc kubenswrapper[4735]: I1122 09:47:15.219866 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wcpq5"] Nov 22 09:47:16 crc kubenswrapper[4735]: I1122 09:47:16.003543 4735 generic.go:334] "Generic (PLEG): container finished" podID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerID="b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1" exitCode=0 Nov 22 09:47:16 crc kubenswrapper[4735]: I1122 09:47:16.003753 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcpq5" event={"ID":"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f","Type":"ContainerDied","Data":"b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1"} Nov 22 09:47:16 crc kubenswrapper[4735]: I1122 09:47:16.004002 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcpq5" event={"ID":"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f","Type":"ContainerStarted","Data":"587f67773ef921a127d0ad67230e81f4a52af491a49d0c6b2ee5b31e78eb8207"} Nov 22 09:47:18 crc kubenswrapper[4735]: I1122 09:47:18.025384 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcpq5" event={"ID":"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f","Type":"ContainerStarted","Data":"83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5"} Nov 22 09:47:20 crc kubenswrapper[4735]: I1122 09:47:20.049892 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-cp7lm_03bbe6ea-a40e-4d51-85fb-d85580cfe27f/cluster-logging-operator/0.log" Nov 22 09:47:20 crc kubenswrapper[4735]: I1122 09:47:20.310718 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-dmrkn_4ea5da34-9c92-4c02-a674-54618b1c5888/collector/0.log" Nov 22 09:47:20 crc kubenswrapper[4735]: I1122 09:47:20.402725 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_77078529-6c53-48d6-9bfc-671f3e2eba43/loki-compactor/0.log" Nov 22 09:47:20 crc kubenswrapper[4735]: I1122 09:47:20.568995 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-hvxhx_2bab3bec-f764-4c68-a487-6d7dbc40e258/loki-distributor/0.log" Nov 22 09:47:20 crc kubenswrapper[4735]: I1122 09:47:20.621454 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6f9c8b877-5dtcw_5caa9595-315a-40a7-a46f-5e6d4b13c320/opa/0.log" Nov 22 09:47:20 crc kubenswrapper[4735]: I1122 09:47:20.647931 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6f9c8b877-5dtcw_5caa9595-315a-40a7-a46f-5e6d4b13c320/gateway/0.log" Nov 22 09:47:20 crc kubenswrapper[4735]: I1122 09:47:20.843029 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6f9c8b877-jv5cr_3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0/gateway/0.log" Nov 22 09:47:20 crc kubenswrapper[4735]: I1122 09:47:20.843199 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6f9c8b877-jv5cr_3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0/opa/0.log" Nov 22 09:47:21 crc kubenswrapper[4735]: I1122 09:47:21.038951 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_cd0c04d0-3cae-4611-9cdb-57e95cbb3a5f/loki-index-gateway/0.log" Nov 22 09:47:21 crc kubenswrapper[4735]: I1122 09:47:21.057827 4735 generic.go:334] "Generic (PLEG): container finished" podID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerID="83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5" exitCode=0 Nov 22 09:47:21 crc kubenswrapper[4735]: I1122 09:47:21.057873 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcpq5" event={"ID":"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f","Type":"ContainerDied","Data":"83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5"} Nov 22 09:47:21 crc kubenswrapper[4735]: I1122 09:47:21.160882 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_d3b1eee2-6b00-45b3-95bd-9e3005004633/loki-ingester/0.log" Nov 22 09:47:21 crc kubenswrapper[4735]: I1122 09:47:21.331381 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-zhskj_35c00115-3277-4e70-ae42-9a13e4a4459e/loki-querier/0.log" Nov 22 09:47:21 crc kubenswrapper[4735]: I1122 09:47:21.415909 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-hxcvg_1b08c63f-3828-4fd2-95a8-7ef002d5fdbd/loki-query-frontend/0.log" Nov 22 09:47:22 crc kubenswrapper[4735]: I1122 09:47:22.071331 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcpq5" event={"ID":"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f","Type":"ContainerStarted","Data":"11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6"} Nov 22 09:47:22 crc kubenswrapper[4735]: I1122 09:47:22.096061 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wcpq5" podStartSLOduration=3.612020669 podStartE2EDuration="9.096044113s" podCreationTimestamp="2025-11-22 09:47:13 +0000 UTC" firstStartedPulling="2025-11-22 09:47:16.005533832 +0000 UTC m=+6257.609872437" lastFinishedPulling="2025-11-22 09:47:21.489557276 +0000 UTC m=+6263.093895881" observedRunningTime="2025-11-22 09:47:22.090481491 +0000 UTC m=+6263.694820106" watchObservedRunningTime="2025-11-22 09:47:22.096044113 +0000 UTC m=+6263.700382718" Nov 22 09:47:22 crc kubenswrapper[4735]: I1122 09:47:22.574810 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-szfj5" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="registry-server" probeResult="failure" output=< Nov 22 09:47:22 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:47:22 crc kubenswrapper[4735]: > Nov 22 09:47:24 crc kubenswrapper[4735]: I1122 09:47:24.245817 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:24 crc kubenswrapper[4735]: I1122 09:47:24.247267 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:25 crc kubenswrapper[4735]: I1122 09:47:25.302504 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-wcpq5" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="registry-server" probeResult="failure" output=< Nov 22 09:47:25 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:47:25 crc kubenswrapper[4735]: > Nov 22 09:47:32 crc kubenswrapper[4735]: I1122 09:47:32.570151 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-szfj5" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="registry-server" probeResult="failure" output=< Nov 22 09:47:32 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:47:32 crc kubenswrapper[4735]: > Nov 22 09:47:34 crc kubenswrapper[4735]: I1122 09:47:34.311768 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:34 crc kubenswrapper[4735]: I1122 09:47:34.384732 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:34 crc kubenswrapper[4735]: I1122 09:47:34.556682 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wcpq5"] Nov 22 09:47:35 crc kubenswrapper[4735]: I1122 09:47:35.919245 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-2gwq7_eddaaabc-450b-4c69-9e29-35f554b01893/kube-rbac-proxy/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.051658 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-2gwq7_eddaaabc-450b-4c69-9e29-35f554b01893/controller/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.068219 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-frr-files/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.279618 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wcpq5" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="registry-server" containerID="cri-o://11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6" gracePeriod=2 Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.319692 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-frr-files/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.358592 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-metrics/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.394088 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-reloader/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.407735 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-reloader/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.662854 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-reloader/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.670689 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-frr-files/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.708101 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-metrics/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.735173 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-metrics/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.787878 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.917380 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-frr-files/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.927173 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-utilities\") pod \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.927236 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-catalog-content\") pod \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.927369 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rghhb\" (UniqueName: \"kubernetes.io/projected/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-kube-api-access-rghhb\") pod \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\" (UID: \"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f\") " Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.928205 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-utilities" (OuterVolumeSpecName: "utilities") pod "e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" (UID: "e135838c-ef83-4c73-b0ae-b3bd4fd9b71f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.963969 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/controller/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.967451 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" (UID: "e135838c-ef83-4c73-b0ae-b3bd4fd9b71f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.967684 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-kube-api-access-rghhb" (OuterVolumeSpecName: "kube-api-access-rghhb") pod "e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" (UID: "e135838c-ef83-4c73-b0ae-b3bd4fd9b71f"). InnerVolumeSpecName "kube-api-access-rghhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.977562 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-reloader/0.log" Nov 22 09:47:36 crc kubenswrapper[4735]: I1122 09:47:36.996615 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/cp-metrics/0.log" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.030231 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.030262 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.030272 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rghhb\" (UniqueName: \"kubernetes.io/projected/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f-kube-api-access-rghhb\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.179553 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/frr-metrics/0.log" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.207487 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/kube-rbac-proxy/0.log" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.259088 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/kube-rbac-proxy-frr/0.log" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.295933 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcpq5" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.295931 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcpq5" event={"ID":"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f","Type":"ContainerDied","Data":"11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6"} Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.296079 4735 scope.go:117] "RemoveContainer" containerID="11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.295935 4735 generic.go:334] "Generic (PLEG): container finished" podID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerID="11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6" exitCode=0 Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.296200 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcpq5" event={"ID":"e135838c-ef83-4c73-b0ae-b3bd4fd9b71f","Type":"ContainerDied","Data":"587f67773ef921a127d0ad67230e81f4a52af491a49d0c6b2ee5b31e78eb8207"} Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.327813 4735 scope.go:117] "RemoveContainer" containerID="83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.347293 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wcpq5"] Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.364137 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wcpq5"] Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.365714 4735 scope.go:117] "RemoveContainer" containerID="b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.436018 4735 scope.go:117] "RemoveContainer" containerID="11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6" Nov 22 09:47:37 crc kubenswrapper[4735]: E1122 09:47:37.439670 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6\": container with ID starting with 11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6 not found: ID does not exist" containerID="11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.439713 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6"} err="failed to get container status \"11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6\": rpc error: code = NotFound desc = could not find container \"11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6\": container with ID starting with 11a28f644b3aee0fe6c3106ec64f61beb4b306de3d1e4e4c60c0477fe51eaac6 not found: ID does not exist" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.439741 4735 scope.go:117] "RemoveContainer" containerID="83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5" Nov 22 09:47:37 crc kubenswrapper[4735]: E1122 09:47:37.442019 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5\": container with ID starting with 83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5 not found: ID does not exist" containerID="83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.442079 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5"} err="failed to get container status \"83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5\": rpc error: code = NotFound desc = could not find container \"83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5\": container with ID starting with 83975604baafd92d280ef50ff53626495c7f7f18bf05e185ce6427df725273d5 not found: ID does not exist" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.442111 4735 scope.go:117] "RemoveContainer" containerID="b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1" Nov 22 09:47:37 crc kubenswrapper[4735]: E1122 09:47:37.442501 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1\": container with ID starting with b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1 not found: ID does not exist" containerID="b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.442543 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1"} err="failed to get container status \"b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1\": rpc error: code = NotFound desc = could not find container \"b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1\": container with ID starting with b8533932fb49f3e71e50c84fdc105a71e3ce84dc2e629b0490f01f22d6a2ddc1 not found: ID does not exist" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.489871 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-wzm9n_84fe49a3-baa7-45bf-aa7b-b92e55e61a1e/frr-k8s-webhook-server/0.log" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.502251 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/reloader/0.log" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.684562 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-f4fbf457b-9k6b5_94a2a8b2-e02e-409a-9fa2-06016f577fb1/manager/0.log" Nov 22 09:47:37 crc kubenswrapper[4735]: I1122 09:47:37.909214 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-b67588696-x9fc4_61883f69-88c2-40a4-b62b-f258bbace22f/webhook-server/0.log" Nov 22 09:47:38 crc kubenswrapper[4735]: I1122 09:47:38.047910 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-klgsh_e40c967d-3d49-4b06-8789-9b1abe501e80/kube-rbac-proxy/0.log" Nov 22 09:47:38 crc kubenswrapper[4735]: I1122 09:47:38.739375 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-klgsh_e40c967d-3d49-4b06-8789-9b1abe501e80/speaker/0.log" Nov 22 09:47:39 crc kubenswrapper[4735]: I1122 09:47:39.301866 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" path="/var/lib/kubelet/pods/e135838c-ef83-4c73-b0ae-b3bd4fd9b71f/volumes" Nov 22 09:47:39 crc kubenswrapper[4735]: I1122 09:47:39.337921 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkh26_2719cd0e-b5e9-4df2-9165-b9bb102b1b70/frr/0.log" Nov 22 09:47:42 crc kubenswrapper[4735]: I1122 09:47:42.594494 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-szfj5" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="registry-server" probeResult="failure" output=< Nov 22 09:47:42 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:47:42 crc kubenswrapper[4735]: > Nov 22 09:47:46 crc kubenswrapper[4735]: I1122 09:47:46.131104 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:47:46 crc kubenswrapper[4735]: I1122 09:47:46.131726 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:47:51 crc kubenswrapper[4735]: I1122 09:47:51.570160 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:51 crc kubenswrapper[4735]: I1122 09:47:51.659536 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:51 crc kubenswrapper[4735]: I1122 09:47:51.816451 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-szfj5"] Nov 22 09:47:51 crc kubenswrapper[4735]: I1122 09:47:51.853207 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6_467ea8b5-523e-40a4-8cd2-9f20f203a975/util/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.035838 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6_467ea8b5-523e-40a4-8cd2-9f20f203a975/util/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.071161 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6_467ea8b5-523e-40a4-8cd2-9f20f203a975/pull/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.090844 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6_467ea8b5-523e-40a4-8cd2-9f20f203a975/pull/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.274074 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6_467ea8b5-523e-40a4-8cd2-9f20f203a975/util/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.295710 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6_467ea8b5-523e-40a4-8cd2-9f20f203a975/pull/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.386985 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb82zkc6_467ea8b5-523e-40a4-8cd2-9f20f203a975/extract/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.484202 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9_bd3ee390-fac8-4865-9be5-469cc9c2d791/util/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.695499 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9_bd3ee390-fac8-4865-9be5-469cc9c2d791/util/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.730665 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9_bd3ee390-fac8-4865-9be5-469cc9c2d791/pull/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.742741 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9_bd3ee390-fac8-4865-9be5-469cc9c2d791/pull/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.887402 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9_bd3ee390-fac8-4865-9be5-469cc9c2d791/pull/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.889233 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9_bd3ee390-fac8-4865-9be5-469cc9c2d791/util/0.log" Nov 22 09:47:52 crc kubenswrapper[4735]: I1122 09:47:52.940897 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejf7p9_bd3ee390-fac8-4865-9be5-469cc9c2d791/extract/0.log" Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.149330 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8_b6c8e68c-78c8-4d48-be93-a5b5d23204e0/util/0.log" Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.305830 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8_b6c8e68c-78c8-4d48-be93-a5b5d23204e0/pull/0.log" Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.346057 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8_b6c8e68c-78c8-4d48-be93-a5b5d23204e0/util/0.log" Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.351077 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8_b6c8e68c-78c8-4d48-be93-a5b5d23204e0/pull/0.log" Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.494262 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-szfj5" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="registry-server" containerID="cri-o://f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39" gracePeriod=2 Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.515199 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8_b6c8e68c-78c8-4d48-be93-a5b5d23204e0/util/0.log" Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.572112 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8_b6c8e68c-78c8-4d48-be93-a5b5d23204e0/pull/0.log" Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.602020 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921055jt8_b6c8e68c-78c8-4d48-be93-a5b5d23204e0/extract/0.log" Nov 22 09:47:53 crc kubenswrapper[4735]: I1122 09:47:53.773937 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k_73825640-996d-43e1-8e5d-e5ed2a514896/util/0.log" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.139796 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k_73825640-996d-43e1-8e5d-e5ed2a514896/util/0.log" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.140269 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k_73825640-996d-43e1-8e5d-e5ed2a514896/pull/0.log" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.262411 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.365748 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-catalog-content\") pod \"3b630869-ac40-452b-b1af-0e0b7c27f717\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.365902 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-utilities\") pod \"3b630869-ac40-452b-b1af-0e0b7c27f717\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.365952 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzsk8\" (UniqueName: \"kubernetes.io/projected/3b630869-ac40-452b-b1af-0e0b7c27f717-kube-api-access-jzsk8\") pod \"3b630869-ac40-452b-b1af-0e0b7c27f717\" (UID: \"3b630869-ac40-452b-b1af-0e0b7c27f717\") " Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.380493 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k_73825640-996d-43e1-8e5d-e5ed2a514896/pull/0.log" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.380911 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-utilities" (OuterVolumeSpecName: "utilities") pod "3b630869-ac40-452b-b1af-0e0b7c27f717" (UID: "3b630869-ac40-452b-b1af-0e0b7c27f717"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.389693 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b630869-ac40-452b-b1af-0e0b7c27f717-kube-api-access-jzsk8" (OuterVolumeSpecName: "kube-api-access-jzsk8") pod "3b630869-ac40-452b-b1af-0e0b7c27f717" (UID: "3b630869-ac40-452b-b1af-0e0b7c27f717"). InnerVolumeSpecName "kube-api-access-jzsk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.468752 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.468782 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzsk8\" (UniqueName: \"kubernetes.io/projected/3b630869-ac40-452b-b1af-0e0b7c27f717-kube-api-access-jzsk8\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.472440 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b630869-ac40-452b-b1af-0e0b7c27f717" (UID: "3b630869-ac40-452b-b1af-0e0b7c27f717"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.473167 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k_73825640-996d-43e1-8e5d-e5ed2a514896/util/0.log" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.504944 4735 generic.go:334] "Generic (PLEG): container finished" podID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerID="f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39" exitCode=0 Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.504995 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-szfj5" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.505008 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szfj5" event={"ID":"3b630869-ac40-452b-b1af-0e0b7c27f717","Type":"ContainerDied","Data":"f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39"} Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.505332 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szfj5" event={"ID":"3b630869-ac40-452b-b1af-0e0b7c27f717","Type":"ContainerDied","Data":"05cf25a3b5309b18a92128656fa96431c5a9875db259cab4aeb29c9b3b55a6eb"} Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.505358 4735 scope.go:117] "RemoveContainer" containerID="f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.520231 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k_73825640-996d-43e1-8e5d-e5ed2a514896/pull/0.log" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.570509 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b630869-ac40-452b-b1af-0e0b7c27f717-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.694268 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4fxx_c6dab36d-6121-4c4b-b350-7594bc1c0133/extract-utilities/0.log" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.909806 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fpnc5k_73825640-996d-43e1-8e5d-e5ed2a514896/extract/0.log" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.924172 4735 scope.go:117] "RemoveContainer" containerID="19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.944187 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-szfj5"] Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.954778 4735 scope.go:117] "RemoveContainer" containerID="7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865" Nov 22 09:47:54 crc kubenswrapper[4735]: I1122 09:47:54.966363 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-szfj5"] Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.025894 4735 scope.go:117] "RemoveContainer" containerID="f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39" Nov 22 09:47:55 crc kubenswrapper[4735]: E1122 09:47:55.027967 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39\": container with ID starting with f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39 not found: ID does not exist" containerID="f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.028102 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39"} err="failed to get container status \"f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39\": rpc error: code = NotFound desc = could not find container \"f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39\": container with ID starting with f39985261f3ebf8f2deb1de47805436460ebb26daf3a222636dc7441bc654e39 not found: ID does not exist" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.028204 4735 scope.go:117] "RemoveContainer" containerID="19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf" Nov 22 09:47:55 crc kubenswrapper[4735]: E1122 09:47:55.028662 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf\": container with ID starting with 19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf not found: ID does not exist" containerID="19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.028702 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf"} err="failed to get container status \"19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf\": rpc error: code = NotFound desc = could not find container \"19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf\": container with ID starting with 19457ee5315572f50e95402d716ca2899cee8e6b42938410490b54717a499baf not found: ID does not exist" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.028731 4735 scope.go:117] "RemoveContainer" containerID="7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865" Nov 22 09:47:55 crc kubenswrapper[4735]: E1122 09:47:55.029268 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865\": container with ID starting with 7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865 not found: ID does not exist" containerID="7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.029288 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865"} err="failed to get container status \"7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865\": rpc error: code = NotFound desc = could not find container \"7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865\": container with ID starting with 7e985b40b8f909dc36375bb55b51a8e76e2db0d36ad6592c26ad655f5401d865 not found: ID does not exist" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.213930 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4fxx_c6dab36d-6121-4c4b-b350-7594bc1c0133/extract-utilities/0.log" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.226564 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4fxx_c6dab36d-6121-4c4b-b350-7594bc1c0133/extract-content/0.log" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.263864 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4fxx_c6dab36d-6121-4c4b-b350-7594bc1c0133/extract-content/0.log" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.277750 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" path="/var/lib/kubelet/pods/3b630869-ac40-452b-b1af-0e0b7c27f717/volumes" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.392066 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4fxx_c6dab36d-6121-4c4b-b350-7594bc1c0133/extract-utilities/0.log" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.427614 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4fxx_c6dab36d-6121-4c4b-b350-7594bc1c0133/extract-content/0.log" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.725622 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p8qf6_70cac5fe-7dac-4acc-87a0-c4194fdae63c/extract-utilities/0.log" Nov 22 09:47:55 crc kubenswrapper[4735]: I1122 09:47:55.952595 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p8qf6_70cac5fe-7dac-4acc-87a0-c4194fdae63c/extract-utilities/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.017640 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p8qf6_70cac5fe-7dac-4acc-87a0-c4194fdae63c/extract-content/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.026037 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p8qf6_70cac5fe-7dac-4acc-87a0-c4194fdae63c/extract-content/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.198208 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p8qf6_70cac5fe-7dac-4acc-87a0-c4194fdae63c/extract-content/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.214346 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4fxx_c6dab36d-6121-4c4b-b350-7594bc1c0133/registry-server/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.230676 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p8qf6_70cac5fe-7dac-4acc-87a0-c4194fdae63c/extract-utilities/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.433496 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59_bb5e21db-7091-4338-8ea2-1a1f3d861984/util/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.585928 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59_bb5e21db-7091-4338-8ea2-1a1f3d861984/pull/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.604668 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59_bb5e21db-7091-4338-8ea2-1a1f3d861984/util/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.640336 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59_bb5e21db-7091-4338-8ea2-1a1f3d861984/pull/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.860277 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59_bb5e21db-7091-4338-8ea2-1a1f3d861984/pull/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.900604 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59_bb5e21db-7091-4338-8ea2-1a1f3d861984/util/0.log" Nov 22 09:47:56 crc kubenswrapper[4735]: I1122 09:47:56.920266 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xxr59_bb5e21db-7091-4338-8ea2-1a1f3d861984/extract/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.141715 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p8qf6_70cac5fe-7dac-4acc-87a0-c4194fdae63c/registry-server/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.162054 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qr6cq_f10ff5dd-2bc5-4b07-90d2-174b896191bd/marketplace-operator/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.213776 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5p7s_f1964504-b509-4e0e-b88e-2577618a493a/extract-utilities/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.364595 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5p7s_f1964504-b509-4e0e-b88e-2577618a493a/extract-content/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.409929 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5p7s_f1964504-b509-4e0e-b88e-2577618a493a/extract-content/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.428070 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5p7s_f1964504-b509-4e0e-b88e-2577618a493a/extract-utilities/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.557950 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5p7s_f1964504-b509-4e0e-b88e-2577618a493a/extract-utilities/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.586877 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5p7s_f1964504-b509-4e0e-b88e-2577618a493a/extract-content/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.615992 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2cq5w_76c916a7-a1da-48e5-b7cd-1827b4d42c7a/extract-utilities/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.865903 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5p7s_f1964504-b509-4e0e-b88e-2577618a493a/registry-server/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.973767 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2cq5w_76c916a7-a1da-48e5-b7cd-1827b4d42c7a/extract-content/0.log" Nov 22 09:47:57 crc kubenswrapper[4735]: I1122 09:47:57.994958 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2cq5w_76c916a7-a1da-48e5-b7cd-1827b4d42c7a/extract-utilities/0.log" Nov 22 09:47:58 crc kubenswrapper[4735]: I1122 09:47:58.005428 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2cq5w_76c916a7-a1da-48e5-b7cd-1827b4d42c7a/extract-content/0.log" Nov 22 09:47:58 crc kubenswrapper[4735]: I1122 09:47:58.177513 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2cq5w_76c916a7-a1da-48e5-b7cd-1827b4d42c7a/extract-utilities/0.log" Nov 22 09:47:58 crc kubenswrapper[4735]: I1122 09:47:58.221560 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2cq5w_76c916a7-a1da-48e5-b7cd-1827b4d42c7a/extract-content/0.log" Nov 22 09:47:58 crc kubenswrapper[4735]: I1122 09:47:58.886394 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2cq5w_76c916a7-a1da-48e5-b7cd-1827b4d42c7a/registry-server/0.log" Nov 22 09:48:11 crc kubenswrapper[4735]: I1122 09:48:11.974159 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-sxnpm_eab1ae65-df97-4ef5-9876-dccbac6d5304/prometheus-operator/0.log" Nov 22 09:48:12 crc kubenswrapper[4735]: I1122 09:48:12.163623 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7d649cf67c-4f2nz_73f91b2d-3de6-40da-ac7b-e97c7b63a8ad/prometheus-operator-admission-webhook/0.log" Nov 22 09:48:12 crc kubenswrapper[4735]: I1122 09:48:12.182042 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7d649cf67c-kj579_50c1fd51-0497-44e6-b9d6-52b48e5a4133/prometheus-operator-admission-webhook/0.log" Nov 22 09:48:12 crc kubenswrapper[4735]: I1122 09:48:12.384650 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-5wrtk_93273ee2-9156-4b65-abe9-1e020aa4ea55/operator/0.log" Nov 22 09:48:12 crc kubenswrapper[4735]: I1122 09:48:12.388112 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-2dl4c_36da6203-0695-45d5-b361-01267bd61a1d/observability-ui-dashboards/0.log" Nov 22 09:48:12 crc kubenswrapper[4735]: I1122 09:48:12.582496 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-jq2p9_fc5129c5-6a4c-4d09-807f-1811604f3193/perses-operator/0.log" Nov 22 09:48:16 crc kubenswrapper[4735]: I1122 09:48:16.132164 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:48:16 crc kubenswrapper[4735]: I1122 09:48:16.132669 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:48:26 crc kubenswrapper[4735]: I1122 09:48:26.400925 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5b4b6fd4bd-jc7r4_b0172333-c8a0-4e17-a312-d3362d0fe8da/kube-rbac-proxy/0.log" Nov 22 09:48:26 crc kubenswrapper[4735]: I1122 09:48:26.472010 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5b4b6fd4bd-jc7r4_b0172333-c8a0-4e17-a312-d3362d0fe8da/manager/0.log" Nov 22 09:48:46 crc kubenswrapper[4735]: I1122 09:48:46.133222 4735 patch_prober.go:28] interesting pod/machine-config-daemon-dt2lg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:48:46 crc kubenswrapper[4735]: I1122 09:48:46.133798 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:48:46 crc kubenswrapper[4735]: I1122 09:48:46.133852 4735 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" Nov 22 09:48:46 crc kubenswrapper[4735]: I1122 09:48:46.134827 4735 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827"} pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:48:46 crc kubenswrapper[4735]: I1122 09:48:46.134931 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerName="machine-config-daemon" containerID="cri-o://dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" gracePeriod=600 Nov 22 09:48:46 crc kubenswrapper[4735]: E1122 09:48:46.312905 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:48:47 crc kubenswrapper[4735]: I1122 09:48:47.186642 4735 generic.go:334] "Generic (PLEG): container finished" podID="b7945a41-afa2-43b7-95a2-085e3b62e64b" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" exitCode=0 Nov 22 09:48:47 crc kubenswrapper[4735]: I1122 09:48:47.186706 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerDied","Data":"dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827"} Nov 22 09:48:47 crc kubenswrapper[4735]: I1122 09:48:47.186780 4735 scope.go:117] "RemoveContainer" containerID="76f202ccd08e999858681fe56c8b8a5850e9e3360e4ffa862758f705684ac588" Nov 22 09:48:47 crc kubenswrapper[4735]: I1122 09:48:47.187622 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:48:47 crc kubenswrapper[4735]: E1122 09:48:47.188010 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.739776 4735 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.562425512s: [/var/lib/containers/storage/overlay/71c10dc7d73e7bce0546d44be1f02432688c4d609f318090c65e83cceb36e83a/diff /var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-xl2mv_9124e032-f3c0-46fa-bd17-26ed40bfd57e/manager/0.log]; will not log again for this container unless duration exceeds 3s Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.741091 4735 patch_prober.go:28] interesting pod/logging-loki-gateway-6f9c8b877-jv5cr container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.76:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.741173 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" podUID="3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.76:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.750214 4735 patch_prober.go:28] interesting pod/logging-loki-gateway-6f9c8b877-jv5cr container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.76:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.750290 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-6f9c8b877-jv5cr" podUID="3fc1eb4e-6044-4fe5-a6ad-ad3dc18892a0" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.76:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.750621 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-f4fbf457b-9k6b5" podUID="94a2a8b2-e02e-409a-9fa2-06016f577fb1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.97:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.750685 4735 patch_prober.go:28] interesting pod/logging-loki-gateway-6f9c8b877-5dtcw container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.77:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.750705 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" podUID="5caa9595-315a-40a7-a46f-5e6d4b13c320" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.77:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.750852 4735 patch_prober.go:28] interesting pod/logging-loki-gateway-6f9c8b877-5dtcw container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.77:8081/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.750880 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-6f9c8b877-5dtcw" podUID="5caa9595-315a-40a7-a46f-5e6d4b13c320" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.77:8081/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.751005 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-bkh26" podUID="2719cd0e-b5e9-4df2-9165-b9bb102b1b70" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.751138 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-bkh26" podUID="2719cd0e-b5e9-4df2-9165-b9bb102b1b70" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.751227 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-bkh26" podUID="2719cd0e-b5e9-4df2-9165-b9bb102b1b70" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.751347 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" podUID="84fe49a3-baa7-45bf-aa7b-b92e55e61a1e" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.99:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.753262 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wzm9n" podUID="84fe49a3-baa7-45bf-aa7b-b92e55e61a1e" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.99:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.754170 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-6c7b4b5f48-2gwq7" podUID="eddaaabc-450b-4c69-9e29-35f554b01893" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.100:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.754816 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-6c7b4b5f48-2gwq7" podUID="eddaaabc-450b-4c69-9e29-35f554b01893" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.100:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.757984 4735 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 5.635026962s: [/var/lib/containers/storage/overlay/939c2046bb10e7292490711e6adba9decddcff37e6ac460fe54b0dde73c84aac/diff /var/log/pods/openstack_keystone-5584fd997b-7ztjv_4906aa84-20dc-4096-8d7c-0ed5dd4d9103/keystone-api/0.log]; will not log again for this container unless duration exceeds 2s Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.762080 4735 patch_prober.go:28] interesting pod/nmstate-webhook-6b89b748d8-swwps container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.93:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.762159 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-swwps" podUID="fd488627-bc81-4723-b465-b39ff151d0d6" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.93:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.771874 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-26wft" podUID="24a36a64-01ec-4f38-b118-e1c6097e8965" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.47:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.772025 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" podUID="2dd266b7-8f4d-46b3-b8e7-3511b5404f8d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.772208 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-26vdp" podUID="2dd266b7-8f4d-46b3-b8e7-3511b5404f8d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.772428 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" podUID="5765b990-2731-42c7-bb14-7b149e91555f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.772552 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-dmhxb" podUID="5765b990-2731-42c7-bb14-7b149e91555f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.772653 4735 patch_prober.go:28] interesting pod/thanos-querier-8579bbc54b-z8th6 container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.68:9091/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.772672 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-8579bbc54b-z8th6" podUID="fcf09211-515e-43bd-bc3b-f12305641d84" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.68:9091/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.772747 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" podUID="2038726f-ef9a-47d2-837e-1b81f72c2d39" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.772844 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-7969689c84-zrv98" podUID="2038726f-ef9a-47d2-837e-1b81f72c2d39" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.773554 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" podUID="1c5cb76f-8964-4245-86e5-56b619e06d1b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.773984 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-gkpjb" podUID="1c5cb76f-8964-4245-86e5-56b619e06d1b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.774596 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" podUID="c3ac6a85-2f1e-4906-ba7d-c9415f523fa8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.774629 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-42kvc" podUID="c3ac6a85-2f1e-4906-ba7d-c9415f523fa8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.776213 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" podUID="2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.776499 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-t625d" podUID="2270ad73-e96d-4e1e-bd2c-e6fb254a1bdb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.776723 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" podUID="fc7fded7-7eee-4227-8569-85d07ea4ecd6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.776998 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/manila-operator-controller-manager-58f887965d-nd4wh" podUID="fc7fded7-7eee-4227-8569-85d07ea4ecd6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.779072 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" podUID="61871ccb-bdb6-4222-88d9-d34c22a5797c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.779236 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-pwlw2" podUID="61871ccb-bdb6-4222-88d9-d34c22a5797c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.781095 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" podUID="2c87c2e3-b68e-4f75-a7cf-2c752e7286e8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.784059 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-j5xsn" podUID="2c87c2e3-b68e-4f75-a7cf-2c752e7286e8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.784254 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-klgsh" podUID="e40c967d-3d49-4b06-8789-9b1abe501e80" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.784503 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-klgsh" podUID="e40c967d-3d49-4b06-8789-9b1abe501e80" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.784645 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" podUID="3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.784765 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-vgsgb" podUID="3f4bbfcc-bebe-47e5-a7f3-6d8b6bdf441c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.793089 4735 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" podUID="9124e032-f3c0-46fa-bd17-26ed40bfd57e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.796984 4735 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 3.956001775s: [/var/lib/containers/storage/overlay/5cd4e5457543e93ec1cefc715c9db51346dfe3e4148d5f7ae498e76f6cb00644/diff /var/log/pods/openstack_neutron-7bb54fbc4c-fnr9g_007a6147-937e-485d-9e4c-1a193fdff463/neutron-api/0.log]; will not log again for this container unless duration exceeds 2s Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.797625 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-xl2mv" podUID="9124e032-f3c0-46fa-bd17-26ed40bfd57e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:57 crc kubenswrapper[4735]: I1122 09:48:57.855281 4735 trace.go:236] Trace[1915923034]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-wmm22" (22-Nov-2025 09:48:52.743) (total time: 5102ms): Nov 22 09:48:57 crc kubenswrapper[4735]: Trace[1915923034]: [5.102093584s] [5.102093584s] END Nov 22 09:48:58 crc kubenswrapper[4735]: I1122 09:48:58.264627 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:48:58 crc kubenswrapper[4735]: E1122 09:48:58.265393 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:49:10 crc kubenswrapper[4735]: I1122 09:49:10.266833 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:49:10 crc kubenswrapper[4735]: E1122 09:49:10.268201 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:49:22 crc kubenswrapper[4735]: I1122 09:49:22.264613 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:49:22 crc kubenswrapper[4735]: E1122 09:49:22.266510 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:49:33 crc kubenswrapper[4735]: I1122 09:49:33.263215 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:49:33 crc kubenswrapper[4735]: E1122 09:49:33.263995 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:49:45 crc kubenswrapper[4735]: I1122 09:49:45.263556 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:49:45 crc kubenswrapper[4735]: E1122 09:49:45.264764 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:50:00 crc kubenswrapper[4735]: I1122 09:50:00.264579 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:50:00 crc kubenswrapper[4735]: E1122 09:50:00.265820 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:50:11 crc kubenswrapper[4735]: I1122 09:50:11.264038 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:50:11 crc kubenswrapper[4735]: E1122 09:50:11.265149 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:50:13 crc kubenswrapper[4735]: I1122 09:50:13.352131 4735 scope.go:117] "RemoveContainer" containerID="edab236d01911533c73a2289a0328c4b650ab4bbcf8be85ac0e43bf6be7c1bfb" Nov 22 09:50:23 crc kubenswrapper[4735]: I1122 09:50:23.264235 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:50:23 crc kubenswrapper[4735]: E1122 09:50:23.265699 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:50:24 crc kubenswrapper[4735]: I1122 09:50:24.919174 4735 generic.go:334] "Generic (PLEG): container finished" podID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerID="e52d195a536f7c64458497cd7b12a33af17abdc43c79bccdc97c518b3c2d9c5c" exitCode=0 Nov 22 09:50:24 crc kubenswrapper[4735]: I1122 09:50:24.919266 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xgm58/must-gather-vrnsx" event={"ID":"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0","Type":"ContainerDied","Data":"e52d195a536f7c64458497cd7b12a33af17abdc43c79bccdc97c518b3c2d9c5c"} Nov 22 09:50:24 crc kubenswrapper[4735]: I1122 09:50:24.920255 4735 scope.go:117] "RemoveContainer" containerID="e52d195a536f7c64458497cd7b12a33af17abdc43c79bccdc97c518b3c2d9c5c" Nov 22 09:50:25 crc kubenswrapper[4735]: I1122 09:50:25.101604 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xgm58_must-gather-vrnsx_72f8149b-0bac-4df4-b61b-7a7c5e44e5d0/gather/0.log" Nov 22 09:50:33 crc kubenswrapper[4735]: I1122 09:50:33.527918 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xgm58/must-gather-vrnsx"] Nov 22 09:50:33 crc kubenswrapper[4735]: I1122 09:50:33.528651 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xgm58/must-gather-vrnsx" podUID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerName="copy" containerID="cri-o://124998c2c39ad7430659eecf4bcc3ee5ea6a0b4b4969547a5238877e54d7441c" gracePeriod=2 Nov 22 09:50:33 crc kubenswrapper[4735]: I1122 09:50:33.546734 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xgm58/must-gather-vrnsx"] Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.031392 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xgm58_must-gather-vrnsx_72f8149b-0bac-4df4-b61b-7a7c5e44e5d0/copy/0.log" Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.031997 4735 generic.go:334] "Generic (PLEG): container finished" podID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerID="124998c2c39ad7430659eecf4bcc3ee5ea6a0b4b4969547a5238877e54d7441c" exitCode=143 Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.264772 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:50:34 crc kubenswrapper[4735]: E1122 09:50:34.265378 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.511791 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xgm58_must-gather-vrnsx_72f8149b-0bac-4df4-b61b-7a7c5e44e5d0/copy/0.log" Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.512237 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.558437 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ddbf\" (UniqueName: \"kubernetes.io/projected/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-kube-api-access-2ddbf\") pod \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\" (UID: \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\") " Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.558616 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-must-gather-output\") pod \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\" (UID: \"72f8149b-0bac-4df4-b61b-7a7c5e44e5d0\") " Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.565539 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-kube-api-access-2ddbf" (OuterVolumeSpecName: "kube-api-access-2ddbf") pod "72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" (UID: "72f8149b-0bac-4df4-b61b-7a7c5e44e5d0"). InnerVolumeSpecName "kube-api-access-2ddbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.661703 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ddbf\" (UniqueName: \"kubernetes.io/projected/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-kube-api-access-2ddbf\") on node \"crc\" DevicePath \"\"" Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.761055 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" (UID: "72f8149b-0bac-4df4-b61b-7a7c5e44e5d0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:50:34 crc kubenswrapper[4735]: I1122 09:50:34.765644 4735 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 22 09:50:35 crc kubenswrapper[4735]: I1122 09:50:35.043930 4735 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xgm58_must-gather-vrnsx_72f8149b-0bac-4df4-b61b-7a7c5e44e5d0/copy/0.log" Nov 22 09:50:35 crc kubenswrapper[4735]: I1122 09:50:35.044536 4735 scope.go:117] "RemoveContainer" containerID="124998c2c39ad7430659eecf4bcc3ee5ea6a0b4b4969547a5238877e54d7441c" Nov 22 09:50:35 crc kubenswrapper[4735]: I1122 09:50:35.044576 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xgm58/must-gather-vrnsx" Nov 22 09:50:35 crc kubenswrapper[4735]: I1122 09:50:35.064540 4735 scope.go:117] "RemoveContainer" containerID="e52d195a536f7c64458497cd7b12a33af17abdc43c79bccdc97c518b3c2d9c5c" Nov 22 09:50:35 crc kubenswrapper[4735]: I1122 09:50:35.274794 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" path="/var/lib/kubelet/pods/72f8149b-0bac-4df4-b61b-7a7c5e44e5d0/volumes" Nov 22 09:50:49 crc kubenswrapper[4735]: I1122 09:50:49.274981 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:50:49 crc kubenswrapper[4735]: E1122 09:50:49.275821 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:51:04 crc kubenswrapper[4735]: I1122 09:51:04.263647 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:51:04 crc kubenswrapper[4735]: E1122 09:51:04.264274 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:51:13 crc kubenswrapper[4735]: I1122 09:51:13.446086 4735 scope.go:117] "RemoveContainer" containerID="e07f676fb9ed965488cc541dbb47f298b9a248c5fd204bf66cc0fb3160217e0a" Nov 22 09:51:13 crc kubenswrapper[4735]: I1122 09:51:13.490505 4735 scope.go:117] "RemoveContainer" containerID="4ca5b0cca0066f58a42fe9588104a7fb2f6d9a4addaf5fca99ef7fa65936fbf1" Nov 22 09:51:18 crc kubenswrapper[4735]: I1122 09:51:18.264051 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:51:18 crc kubenswrapper[4735]: E1122 09:51:18.266901 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:51:31 crc kubenswrapper[4735]: I1122 09:51:31.263250 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:51:31 crc kubenswrapper[4735]: E1122 09:51:31.263938 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:51:45 crc kubenswrapper[4735]: I1122 09:51:45.264381 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:51:45 crc kubenswrapper[4735]: E1122 09:51:45.265618 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:52:00 crc kubenswrapper[4735]: I1122 09:52:00.263782 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:52:00 crc kubenswrapper[4735]: E1122 09:52:00.264861 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:52:03 crc kubenswrapper[4735]: I1122 09:52:03.171286 4735 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9v46s container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 09:52:03 crc kubenswrapper[4735]: I1122 09:52:03.171833 4735 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9v46s" podUID="a24168d2-50e3-4ee4-bcbf-7abf4b57152e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 09:52:11 crc kubenswrapper[4735]: I1122 09:52:11.263671 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:52:11 crc kubenswrapper[4735]: E1122 09:52:11.264614 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.840991 4735 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cm48k"] Nov 22 09:52:20 crc kubenswrapper[4735]: E1122 09:52:20.842172 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="extract-content" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842188 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="extract-content" Nov 22 09:52:20 crc kubenswrapper[4735]: E1122 09:52:20.842208 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="registry-server" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842216 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="registry-server" Nov 22 09:52:20 crc kubenswrapper[4735]: E1122 09:52:20.842236 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerName="gather" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842244 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerName="gather" Nov 22 09:52:20 crc kubenswrapper[4735]: E1122 09:52:20.842269 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="extract-content" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842278 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="extract-content" Nov 22 09:52:20 crc kubenswrapper[4735]: E1122 09:52:20.842300 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="extract-utilities" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842308 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="extract-utilities" Nov 22 09:52:20 crc kubenswrapper[4735]: E1122 09:52:20.842344 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="registry-server" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842352 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="registry-server" Nov 22 09:52:20 crc kubenswrapper[4735]: E1122 09:52:20.842380 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerName="copy" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842389 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerName="copy" Nov 22 09:52:20 crc kubenswrapper[4735]: E1122 09:52:20.842403 4735 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="extract-utilities" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842413 4735 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="extract-utilities" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842792 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b630869-ac40-452b-b1af-0e0b7c27f717" containerName="registry-server" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842812 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerName="gather" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842830 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="72f8149b-0bac-4df4-b61b-7a7c5e44e5d0" containerName="copy" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.842884 4735 memory_manager.go:354] "RemoveStaleState removing state" podUID="e135838c-ef83-4c73-b0ae-b3bd4fd9b71f" containerName="registry-server" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.846242 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.938125 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-catalog-content\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.938588 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csq6d\" (UniqueName: \"kubernetes.io/projected/c48f6578-136c-40d8-879a-a4a8fca4e428-kube-api-access-csq6d\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:20 crc kubenswrapper[4735]: I1122 09:52:20.938714 4735 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-utilities\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.041523 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csq6d\" (UniqueName: \"kubernetes.io/projected/c48f6578-136c-40d8-879a-a4a8fca4e428-kube-api-access-csq6d\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.041679 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-utilities\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.041909 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm48k"] Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.042011 4735 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-catalog-content\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.042557 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-utilities\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.042630 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-catalog-content\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.073088 4735 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csq6d\" (UniqueName: \"kubernetes.io/projected/c48f6578-136c-40d8-879a-a4a8fca4e428-kube-api-access-csq6d\") pod \"redhat-marketplace-cm48k\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.171157 4735 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:21 crc kubenswrapper[4735]: I1122 09:52:21.997956 4735 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm48k"] Nov 22 09:52:22 crc kubenswrapper[4735]: I1122 09:52:22.263255 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:52:22 crc kubenswrapper[4735]: E1122 09:52:22.264035 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:52:22 crc kubenswrapper[4735]: I1122 09:52:22.376929 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm48k" event={"ID":"c48f6578-136c-40d8-879a-a4a8fca4e428","Type":"ContainerStarted","Data":"3920a10e10c33419b2d14c49bfa257317f382438012e4c20de2221e3beab6d6b"} Nov 22 09:52:23 crc kubenswrapper[4735]: I1122 09:52:23.391652 4735 generic.go:334] "Generic (PLEG): container finished" podID="c48f6578-136c-40d8-879a-a4a8fca4e428" containerID="e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6" exitCode=0 Nov 22 09:52:23 crc kubenswrapper[4735]: I1122 09:52:23.391735 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm48k" event={"ID":"c48f6578-136c-40d8-879a-a4a8fca4e428","Type":"ContainerDied","Data":"e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6"} Nov 22 09:52:23 crc kubenswrapper[4735]: I1122 09:52:23.396051 4735 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:52:25 crc kubenswrapper[4735]: I1122 09:52:25.416313 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm48k" event={"ID":"c48f6578-136c-40d8-879a-a4a8fca4e428","Type":"ContainerStarted","Data":"4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924"} Nov 22 09:52:27 crc kubenswrapper[4735]: I1122 09:52:27.441147 4735 generic.go:334] "Generic (PLEG): container finished" podID="c48f6578-136c-40d8-879a-a4a8fca4e428" containerID="4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924" exitCode=0 Nov 22 09:52:27 crc kubenswrapper[4735]: I1122 09:52:27.441256 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm48k" event={"ID":"c48f6578-136c-40d8-879a-a4a8fca4e428","Type":"ContainerDied","Data":"4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924"} Nov 22 09:52:29 crc kubenswrapper[4735]: I1122 09:52:29.475890 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm48k" event={"ID":"c48f6578-136c-40d8-879a-a4a8fca4e428","Type":"ContainerStarted","Data":"d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e"} Nov 22 09:52:29 crc kubenswrapper[4735]: I1122 09:52:29.515525 4735 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cm48k" podStartSLOduration=3.823572461 podStartE2EDuration="9.515500533s" podCreationTimestamp="2025-11-22 09:52:20 +0000 UTC" firstStartedPulling="2025-11-22 09:52:23.395759496 +0000 UTC m=+6565.000098111" lastFinishedPulling="2025-11-22 09:52:29.087687578 +0000 UTC m=+6570.692026183" observedRunningTime="2025-11-22 09:52:29.503423307 +0000 UTC m=+6571.107761922" watchObservedRunningTime="2025-11-22 09:52:29.515500533 +0000 UTC m=+6571.119839138" Nov 22 09:52:31 crc kubenswrapper[4735]: I1122 09:52:31.171554 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:31 crc kubenswrapper[4735]: I1122 09:52:31.172439 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:32 crc kubenswrapper[4735]: I1122 09:52:32.226700 4735 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-cm48k" podUID="c48f6578-136c-40d8-879a-a4a8fca4e428" containerName="registry-server" probeResult="failure" output=< Nov 22 09:52:32 crc kubenswrapper[4735]: timeout: failed to connect service ":50051" within 1s Nov 22 09:52:32 crc kubenswrapper[4735]: > Nov 22 09:52:34 crc kubenswrapper[4735]: I1122 09:52:34.264428 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:52:34 crc kubenswrapper[4735]: E1122 09:52:34.265271 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:52:41 crc kubenswrapper[4735]: I1122 09:52:41.278288 4735 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:41 crc kubenswrapper[4735]: I1122 09:52:41.337901 4735 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:41 crc kubenswrapper[4735]: I1122 09:52:41.524746 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm48k"] Nov 22 09:52:42 crc kubenswrapper[4735]: I1122 09:52:42.641391 4735 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cm48k" podUID="c48f6578-136c-40d8-879a-a4a8fca4e428" containerName="registry-server" containerID="cri-o://d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e" gracePeriod=2 Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.164909 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.238201 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-catalog-content\") pod \"c48f6578-136c-40d8-879a-a4a8fca4e428\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.238327 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csq6d\" (UniqueName: \"kubernetes.io/projected/c48f6578-136c-40d8-879a-a4a8fca4e428-kube-api-access-csq6d\") pod \"c48f6578-136c-40d8-879a-a4a8fca4e428\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.238526 4735 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-utilities\") pod \"c48f6578-136c-40d8-879a-a4a8fca4e428\" (UID: \"c48f6578-136c-40d8-879a-a4a8fca4e428\") " Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.240017 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-utilities" (OuterVolumeSpecName: "utilities") pod "c48f6578-136c-40d8-879a-a4a8fca4e428" (UID: "c48f6578-136c-40d8-879a-a4a8fca4e428"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.245936 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48f6578-136c-40d8-879a-a4a8fca4e428-kube-api-access-csq6d" (OuterVolumeSpecName: "kube-api-access-csq6d") pod "c48f6578-136c-40d8-879a-a4a8fca4e428" (UID: "c48f6578-136c-40d8-879a-a4a8fca4e428"). InnerVolumeSpecName "kube-api-access-csq6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.267316 4735 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c48f6578-136c-40d8-879a-a4a8fca4e428" (UID: "c48f6578-136c-40d8-879a-a4a8fca4e428"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.342398 4735 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.342449 4735 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48f6578-136c-40d8-879a-a4a8fca4e428-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.342496 4735 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csq6d\" (UniqueName: \"kubernetes.io/projected/c48f6578-136c-40d8-879a-a4a8fca4e428-kube-api-access-csq6d\") on node \"crc\" DevicePath \"\"" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.656183 4735 generic.go:334] "Generic (PLEG): container finished" podID="c48f6578-136c-40d8-879a-a4a8fca4e428" containerID="d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e" exitCode=0 Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.656235 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm48k" event={"ID":"c48f6578-136c-40d8-879a-a4a8fca4e428","Type":"ContainerDied","Data":"d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e"} Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.656268 4735 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cm48k" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.656657 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm48k" event={"ID":"c48f6578-136c-40d8-879a-a4a8fca4e428","Type":"ContainerDied","Data":"3920a10e10c33419b2d14c49bfa257317f382438012e4c20de2221e3beab6d6b"} Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.656695 4735 scope.go:117] "RemoveContainer" containerID="d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.679109 4735 scope.go:117] "RemoveContainer" containerID="4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.689559 4735 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm48k"] Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.702898 4735 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm48k"] Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.725755 4735 scope.go:117] "RemoveContainer" containerID="e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.782833 4735 scope.go:117] "RemoveContainer" containerID="d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e" Nov 22 09:52:43 crc kubenswrapper[4735]: E1122 09:52:43.783375 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e\": container with ID starting with d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e not found: ID does not exist" containerID="d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.783432 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e"} err="failed to get container status \"d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e\": rpc error: code = NotFound desc = could not find container \"d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e\": container with ID starting with d67e7128d8e0d932cd5d0be14d1e1b68a79a334b8d6d53a9db34f6dbaf4e023e not found: ID does not exist" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.783479 4735 scope.go:117] "RemoveContainer" containerID="4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924" Nov 22 09:52:43 crc kubenswrapper[4735]: E1122 09:52:43.783855 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924\": container with ID starting with 4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924 not found: ID does not exist" containerID="4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.783890 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924"} err="failed to get container status \"4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924\": rpc error: code = NotFound desc = could not find container \"4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924\": container with ID starting with 4167e761d5ad89b0e271b173e64a8e0a0169bc4ddd31ea69b26ed2e75b3c8924 not found: ID does not exist" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.783913 4735 scope.go:117] "RemoveContainer" containerID="e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6" Nov 22 09:52:43 crc kubenswrapper[4735]: E1122 09:52:43.784178 4735 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6\": container with ID starting with e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6 not found: ID does not exist" containerID="e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6" Nov 22 09:52:43 crc kubenswrapper[4735]: I1122 09:52:43.784210 4735 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6"} err="failed to get container status \"e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6\": rpc error: code = NotFound desc = could not find container \"e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6\": container with ID starting with e214b48875629dbd365e4172acd995373202274e3d6ffbdd73e5cae22f85ffe6 not found: ID does not exist" Nov 22 09:52:45 crc kubenswrapper[4735]: I1122 09:52:45.281760 4735 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48f6578-136c-40d8-879a-a4a8fca4e428" path="/var/lib/kubelet/pods/c48f6578-136c-40d8-879a-a4a8fca4e428/volumes" Nov 22 09:52:49 crc kubenswrapper[4735]: I1122 09:52:49.272917 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:52:49 crc kubenswrapper[4735]: E1122 09:52:49.274091 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:53:03 crc kubenswrapper[4735]: I1122 09:53:03.265198 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:53:03 crc kubenswrapper[4735]: E1122 09:53:03.265979 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:53:15 crc kubenswrapper[4735]: I1122 09:53:15.268480 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:53:15 crc kubenswrapper[4735]: E1122 09:53:15.269272 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:53:30 crc kubenswrapper[4735]: I1122 09:53:30.264168 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:53:30 crc kubenswrapper[4735]: E1122 09:53:30.265045 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:53:41 crc kubenswrapper[4735]: I1122 09:53:41.264092 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:53:41 crc kubenswrapper[4735]: E1122 09:53:41.264967 4735 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dt2lg_openshift-machine-config-operator(b7945a41-afa2-43b7-95a2-085e3b62e64b)\"" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" podUID="b7945a41-afa2-43b7-95a2-085e3b62e64b" Nov 22 09:53:54 crc kubenswrapper[4735]: I1122 09:53:54.264068 4735 scope.go:117] "RemoveContainer" containerID="dc9985a010a462eeaf2a0f3e3bec555ff208f4ae89a37570ef7eb40854aae827" Nov 22 09:53:55 crc kubenswrapper[4735]: I1122 09:53:55.534044 4735 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dt2lg" event={"ID":"b7945a41-afa2-43b7-95a2-085e3b62e64b","Type":"ContainerStarted","Data":"4049fa4ab65882ea9f633a71e85fe463c51b1f6677aac5e3044cdec634ef1872"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110304117024436 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110304120017345 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110266527016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110266527015462 5ustar corecore